- abort(WriterCommitMessage[]) - 接口 中的方法org.apache.spark.sql.sources.v2.writer.DataSourceWriter
-
- abort() - 接口 中的方法org.apache.spark.sql.sources.v2.writer.DataWriter
-
Aborts this writer if it is failed.
- abort(long, WriterCommitMessage[]) - 接口 中的方法org.apache.spark.sql.sources.v2.writer.streaming.StreamWriter
-
- abort(WriterCommitMessage[]) - 接口 中的方法org.apache.spark.sql.sources.v2.writer.streaming.StreamWriter
-
- abortJob(JobContext) - 类 中的方法org.apache.spark.internal.io.FileCommitProtocol
-
Aborts a job after the writes fail.
- abortJob(JobContext) - 类 中的方法org.apache.spark.internal.io.HadoopMapReduceCommitProtocol
-
- abortTask(TaskAttemptContext) - 类 中的方法org.apache.spark.internal.io.FileCommitProtocol
-
Aborts a task after the writes have failed.
- abortTask(TaskAttemptContext) - 类 中的方法org.apache.spark.internal.io.HadoopMapReduceCommitProtocol
-
- abs(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Computes the absolute value.
- abs() - 类 中的方法org.apache.spark.sql.types.Decimal
-
- absent() - 类 中的静态方法org.apache.spark.api.java.Optional
-
- AbsoluteError - org.apache.spark.mllib.tree.loss中的类
-
:: DeveloperApi ::
Class for absolute error loss calculation (for regression).
- AbsoluteError() - 类 的构造器org.apache.spark.mllib.tree.loss.AbsoluteError
-
- AbstractLauncher<T extends AbstractLauncher> - org.apache.spark.launcher中的类
-
Base class for launcher implementations.
- accept(Parsers) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaParser
-
- accept(ES, Function1<ES, List<Object>>) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaParser
-
- accept(String, PartialFunction<Object, U>) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaParser
-
- accept(Path) - 类 中的方法org.apache.spark.ml.image.SamplePathFilter
-
- acceptIf(Function1<Object, Object>, Function1<Object, String>) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaParser
-
- acceptMatch(String, PartialFunction<Object, U>) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaParser
-
- acceptSeq(ES, Function1<ES, Iterable<Object>>) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaParser
-
- acceptsType(DataType) - 类 中的方法org.apache.spark.sql.types.ObjectType
-
- accId() - 类 中的方法org.apache.spark.CleanAccum
-
- Accumulable<R,T> - org.apache.spark中的类
-
- Accumulable(R, AccumulableParam<R, T>) - 类 的构造器org.apache.spark.Accumulable
-
已过时。
- accumulable(T, AccumulableParam<T, R>) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
- accumulable(T, String, AccumulableParam<T, R>) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
- accumulable(R, AccumulableParam<R, T>) - 类 中的方法org.apache.spark.SparkContext
-
- accumulable(R, String, AccumulableParam<R, T>) - 类 中的方法org.apache.spark.SparkContext
-
- accumulableCollection(R, Function1<R, Growable<T>>, ClassTag<R>) - 类 中的方法org.apache.spark.SparkContext
-
- AccumulableInfo - org.apache.spark.scheduler中的类
-
:: DeveloperApi ::
Information about an
Accumulable modified during a task or stage.
- AccumulableInfo - org.apache.spark.status.api.v1中的类
-
- accumulableInfoFromJson(JsonAST.JValue) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- accumulableInfoToJson(AccumulableInfo) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- AccumulableParam<R,T> - org.apache.spark中的接口
-
- accumulables() - 类 中的方法org.apache.spark.scheduler.StageInfo
-
Terminal values of accumulables updated during this stage, including all the user-defined
accumulators.
- accumulables() - 类 中的方法org.apache.spark.scheduler.TaskInfo
-
Intermediate updates to accumulables during this task.
- accumulablesToJson(Traversable<AccumulableInfo>) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- Accumulator<T> - org.apache.spark中的类
-
- accumulator(int) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
- accumulator(int, String) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
- accumulator(double) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
- accumulator(double, String) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
- accumulator(T, AccumulatorParam<T>) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
- accumulator(T, String, AccumulatorParam<T>) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
- accumulator(T, AccumulatorParam<T>) - 类 中的方法org.apache.spark.SparkContext
-
- accumulator(T, String, AccumulatorParam<T>) - 类 中的方法org.apache.spark.SparkContext
-
- AccumulatorContext - org.apache.spark.util中的类
-
An internal class used to track accumulators by Spark itself.
- AccumulatorContext() - 类 的构造器org.apache.spark.util.AccumulatorContext
-
- AccumulatorParam<T> - org.apache.spark中的接口
-
- AccumulatorParam.DoubleAccumulatorParam$ - org.apache.spark中的类
-
- AccumulatorParam.FloatAccumulatorParam$ - org.apache.spark中的类
-
- AccumulatorParam.IntAccumulatorParam$ - org.apache.spark中的类
-
- AccumulatorParam.LongAccumulatorParam$ - org.apache.spark中的类
-
- AccumulatorParam.StringAccumulatorParam$ - org.apache.spark中的类
-
- ACCUMULATORS() - 类 中的静态方法org.apache.spark.status.TaskIndexNames
-
- accumulatorUpdates() - 类 中的方法org.apache.spark.status.api.v1.StageData
-
- accumulatorUpdates() - 类 中的方法org.apache.spark.status.api.v1.TaskData
-
- AccumulatorV2<IN,OUT> - org.apache.spark.util中的类
-
The base class for accumulators, that can accumulate inputs of type IN, and produce output of
type OUT.
- AccumulatorV2() - 类 的构造器org.apache.spark.util.AccumulatorV2
-
- accumUpdates() - 类 中的方法org.apache.spark.ExceptionFailure
-
- accumUpdates() - 类 中的方法org.apache.spark.scheduler.SparkListenerExecutorMetricsUpdate
-
- accuracy() - 接口 中的方法org.apache.spark.ml.classification.LogisticRegressionSummary
-
Returns accuracy.
- accuracy() - 类 中的方法org.apache.spark.mllib.evaluation.MulticlassMetrics
-
Returns accuracy
(equals to the total number of correctly classified instances
out of the total number of instances.)
- accuracy() - 类 中的方法org.apache.spark.mllib.evaluation.MultilabelMetrics
-
Returns accuracy
- acos(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
- acos(String) - 类 中的静态方法org.apache.spark.sql.functions
-
- active() - 类 中的方法org.apache.spark.sql.streaming.StreamingQueryManager
-
Returns a list of active queries associated with this SQLContext
- active() - 类 中的方法org.apache.spark.streaming.scheduler.ReceiverInfo
-
- ACTIVE() - 类 中的静态方法org.apache.spark.streaming.scheduler.ReceiverState
-
- activeStages() - 类 中的方法org.apache.spark.status.LiveJob
-
- activeTasks() - 类 中的方法org.apache.spark.status.api.v1.ExecutorSummary
-
- activeTasks() - 类 中的方法org.apache.spark.status.LiveExecutor
-
- activeTasks() - 类 中的方法org.apache.spark.status.LiveJob
-
- activeTasks() - 类 中的方法org.apache.spark.status.LiveStage
-
- add(T) - 类 中的方法org.apache.spark.Accumulable
-
已过时。
Add more data to this accumulator / accumulable
- add(T) - 类 中的静态方法org.apache.spark.Accumulator
-
已过时。
- add(Vector) - 类 中的方法org.apache.spark.ml.clustering.ExpectationAggregator
-
Add a new training instance to this ExpectationAggregator, update the weights,
means and covariances for each distributions, and update the log likelihood.
- add(AFTPoint) - 类 中的方法org.apache.spark.ml.regression.AFTAggregator
-
Add a new training data to this AFTAggregator, and update the loss and gradient
of the objective function.
- add(double[], MultivariateGaussian[], ExpectationSum, Vector<Object>) - 类 中的静态方法org.apache.spark.mllib.clustering.ExpectationSum
-
- add(Vector) - 类 中的方法org.apache.spark.mllib.feature.IDF.DocumentFrequencyAggregator
-
Adds a new document.
- add(BlockMatrix) - 类 中的方法org.apache.spark.mllib.linalg.distributed.BlockMatrix
-
Adds the given block matrix other to this block matrix: this + other.
- add(Vector) - 类 中的方法org.apache.spark.mllib.stat.MultivariateOnlineSummarizer
-
Add a new sample to this summarizer, and update the statistical summary.
- add(StructField) - 类 中的方法org.apache.spark.sql.types.StructType
-
- add(String, DataType) - 类 中的方法org.apache.spark.sql.types.StructType
-
Creates a new
StructType by adding a new nullable field with no metadata.
- add(String, DataType, boolean) - 类 中的方法org.apache.spark.sql.types.StructType
-
Creates a new
StructType by adding a new field with no metadata.
- add(String, DataType, boolean, Metadata) - 类 中的方法org.apache.spark.sql.types.StructType
-
Creates a new
StructType by adding a new field and specifying metadata.
- add(String, DataType, boolean, String) - 类 中的方法org.apache.spark.sql.types.StructType
-
Creates a new
StructType by adding a new field and specifying metadata.
- add(String, String) - 类 中的方法org.apache.spark.sql.types.StructType
-
Creates a new
StructType by adding a new nullable field with no metadata where the
dataType is specified as a String.
- add(String, String, boolean) - 类 中的方法org.apache.spark.sql.types.StructType
-
Creates a new
StructType by adding a new field with no metadata where the
dataType is specified as a String.
- add(String, String, boolean, Metadata) - 类 中的方法org.apache.spark.sql.types.StructType
-
Creates a new
StructType by adding a new field and specifying metadata where the
dataType is specified as a String.
- add(String, String, boolean, String) - 类 中的方法org.apache.spark.sql.types.StructType
-
Creates a new
StructType by adding a new field and specifying metadata where the
dataType is specified as a String.
- add(long, long) - 类 中的静态方法org.apache.spark.streaming.util.RawTextHelper
-
- add(IN) - 类 中的方法org.apache.spark.util.AccumulatorV2
-
Takes the inputs and accumulates.
- add(T) - 类 中的方法org.apache.spark.util.CollectionAccumulator
-
- add(Double) - 类 中的方法org.apache.spark.util.DoubleAccumulator
-
Adds v to the accumulator, i.e. increment sum by v and count by 1.
- add(double) - 类 中的方法org.apache.spark.util.DoubleAccumulator
-
Adds v to the accumulator, i.e. increment sum by v and count by 1.
- add(T) - 类 中的方法org.apache.spark.util.LegacyAccumulatorWrapper
-
- add(Long) - 类 中的方法org.apache.spark.util.LongAccumulator
-
Adds v to the accumulator, i.e. increment sum by v and count by 1.
- add(long) - 类 中的方法org.apache.spark.util.LongAccumulator
-
Adds v to the accumulator, i.e. increment sum by v and count by 1.
- add(Object) - 类 中的方法org.apache.spark.util.sketch.CountMinSketch
-
Increments item's count by one.
- add(Object, long) - 类 中的方法org.apache.spark.util.sketch.CountMinSketch
-
Increments item's count by count.
- add_months(Column, int) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the date that is numMonths after startDate.
- addAccumulator(R, T) - 接口 中的方法org.apache.spark.AccumulableParam
-
已过时。
Add additional data to the accumulator value.
- addAccumulator(T, T) - 接口 中的方法org.apache.spark.AccumulatorParam
-
已过时。
- addAppArgs(String...) - 类 中的方法org.apache.spark.launcher.AbstractLauncher
-
Adds command line arguments for the application.
- addAppArgs(String...) - 类 中的方法org.apache.spark.launcher.SparkLauncher
-
- addBinary(byte[]) - 类 中的方法org.apache.spark.util.sketch.CountMinSketch
-
Increments item's count by one.
- addBinary(byte[], long) - 类 中的方法org.apache.spark.util.sketch.CountMinSketch
-
Increments item's count by count.
- addFile(String) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
Add a file to be downloaded with this Spark job on every node.
- addFile(String, boolean) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
Add a file to be downloaded with this Spark job on every node.
- addFile(String) - 类 中的方法org.apache.spark.launcher.AbstractLauncher
-
Adds a file to be submitted with the application.
- addFile(String) - 类 中的方法org.apache.spark.launcher.SparkLauncher
-
- addFile(String) - 类 中的方法org.apache.spark.SparkContext
-
Add a file to be downloaded with this Spark job on every node.
- addFile(String, boolean) - 类 中的方法org.apache.spark.SparkContext
-
Add a file to be downloaded with this Spark job on every node.
- addFilters(Seq<ServletContextHandler>, SparkConf) - 类 中的静态方法org.apache.spark.ui.JettyUtils
-
Add filters, if any, to the given list of ServletContextHandlers
- addGrid(Param<T>, Iterable<T>) - 类 中的方法org.apache.spark.ml.tuning.ParamGridBuilder
-
Adds a param with multiple values (overwrites if the input param exists).
- addGrid(DoubleParam, double[]) - 类 中的方法org.apache.spark.ml.tuning.ParamGridBuilder
-
Adds a double param with multiple values.
- addGrid(IntParam, int[]) - 类 中的方法org.apache.spark.ml.tuning.ParamGridBuilder
-
Adds an int param with multiple values.
- addGrid(FloatParam, float[]) - 类 中的方法org.apache.spark.ml.tuning.ParamGridBuilder
-
Adds a float param with multiple values.
- addGrid(LongParam, long[]) - 类 中的方法org.apache.spark.ml.tuning.ParamGridBuilder
-
Adds a long param with multiple values.
- addGrid(BooleanParam) - 类 中的方法org.apache.spark.ml.tuning.ParamGridBuilder
-
Adds a boolean param with true and false.
- addInPlace(R, R) - 接口 中的方法org.apache.spark.AccumulableParam
-
已过时。
Merge two accumulated values together.
- addInPlace(double, double) - 类 中的方法org.apache.spark.AccumulatorParam.DoubleAccumulatorParam$
-
已过时。
- addInPlace(float, float) - 类 中的方法org.apache.spark.AccumulatorParam.FloatAccumulatorParam$
-
已过时。
- addInPlace(int, int) - 类 中的方法org.apache.spark.AccumulatorParam.IntAccumulatorParam$
-
已过时。
- addInPlace(long, long) - 类 中的方法org.apache.spark.AccumulatorParam.LongAccumulatorParam$
-
已过时。
- addInPlace(String, String) - 类 中的方法org.apache.spark.AccumulatorParam.StringAccumulatorParam$
-
已过时。
- addJar(String) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
Adds a JAR dependency for all tasks to be executed on this SparkContext in the future.
- addJar(String) - 类 中的方法org.apache.spark.launcher.AbstractLauncher
-
Adds a jar file to be submitted with the application.
- addJar(String) - 类 中的方法org.apache.spark.launcher.SparkLauncher
-
- addJar(String) - 类 中的方法org.apache.spark.SparkContext
-
Adds a JAR dependency for all tasks to be executed on this SparkContext in the future.
- addJar(String) - 类 中的方法org.apache.spark.sql.hive.HiveSessionResourceLoader
-
- addListener(SparkAppHandle.Listener) - 接口 中的方法org.apache.spark.launcher.SparkAppHandle
-
Adds a listener to be notified of changes to the handle's information.
- addListener(L) - 类 中的静态方法org.apache.spark.scheduler.AsyncEventQueue
-
- addListener(StreamingQueryListener) - 类 中的方法org.apache.spark.sql.streaming.StreamingQueryManager
-
- addLocalConfiguration(String, int, int, int, JobConf) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
Add Hadoop configuration specific to a single partition and attempt.
- addLong(long) - 类 中的方法org.apache.spark.util.sketch.CountMinSketch
-
Increments item's count by one.
- addLong(long, long) - 类 中的方法org.apache.spark.util.sketch.CountMinSketch
-
Increments item's count by count.
- addMapOutput(int, MapStatus) - 类 中的方法org.apache.spark.ShuffleStatus
-
Register a map output.
- addMetrics(TaskMetrics, TaskMetrics) - 类 中的静态方法org.apache.spark.status.LiveEntityHelpers
-
Add m2 values to m1.
- addPartition(LiveRDDPartition) - 类 中的方法org.apache.spark.status.RDDPartitionSeq
-
- addPartToPGroup(Partition, PartitionGroup) - 类 中的方法org.apache.spark.rdd.DefaultPartitionCoalescer
-
- addPyFile(String) - 类 中的方法org.apache.spark.launcher.AbstractLauncher
-
Adds a python file / zip / egg to be submitted with the application.
- addPyFile(String) - 类 中的方法org.apache.spark.launcher.SparkLauncher
-
- address() - 类 中的方法org.apache.spark.status.api.v1.RDDDataDistribution
-
- addShutdownHook(Function0<BoxedUnit>) - 类 中的静态方法org.apache.spark.util.ShutdownHookManager
-
Adds a shutdown hook with default priority.
- addShutdownHook(int, Function0<BoxedUnit>) - 类 中的静态方法org.apache.spark.util.ShutdownHookManager
-
Adds a shutdown hook with the given priority.
- addSparkArg(String) - 类 中的方法org.apache.spark.launcher.AbstractLauncher
-
Adds a no-value argument to the Spark invocation.
- addSparkArg(String, String) - 类 中的方法org.apache.spark.launcher.AbstractLauncher
-
Adds an argument with a value to the Spark invocation.
- addSparkArg(String) - 类 中的方法org.apache.spark.launcher.SparkLauncher
-
- addSparkArg(String, String) - 类 中的方法org.apache.spark.launcher.SparkLauncher
-
- addSparkListener(SparkListenerInterface) - 类 中的方法org.apache.spark.SparkContext
-
:: DeveloperApi ::
Register a listener to receive up-calls from events that happen during execution.
- addStreamingListener(StreamingListener) - 类 中的方法org.apache.spark.streaming.api.java.JavaStreamingContext
-
- addStreamingListener(StreamingListener) - 类 中的方法org.apache.spark.streaming.StreamingContext
-
- addString(StringBuilder, String, String, String) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- addString(StringBuilder, String) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- addString(StringBuilder) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- addString(String) - 类 中的方法org.apache.spark.util.sketch.CountMinSketch
-
Increments item's count by one.
- addString(String, long) - 类 中的方法org.apache.spark.util.sketch.CountMinSketch
-
Increments item's count by count.
- addSuppressed(Throwable) - 异常错误 中的静态方法org.apache.spark.sql.AnalysisException
-
- addTaskCompletionListener(TaskCompletionListener) - 类 中的方法org.apache.spark.TaskContext
-
Adds a (Java friendly) listener to be executed on task completion.
- addTaskCompletionListener(Function1<TaskContext, BoxedUnit>) - 类 中的方法org.apache.spark.TaskContext
-
Adds a listener in the form of a Scala closure to be executed on task completion.
- addTaskFailureListener(TaskFailureListener) - 类 中的方法org.apache.spark.TaskContext
-
Adds a listener to be executed on task failure.
- addTaskFailureListener(Function2<TaskContext, Throwable, BoxedUnit>) - 类 中的方法org.apache.spark.TaskContext
-
Adds a listener to be executed on task failure.
- addTime() - 类 中的方法org.apache.spark.status.api.v1.ExecutorSummary
-
- addTime() - 类 中的方法org.apache.spark.status.LiveExecutor
-
- AddWebUIFilter(String, Map<String, String>, String) - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.AddWebUIFilter
-
- AddWebUIFilter$() - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.AddWebUIFilter$
-
- AFTAggregator - org.apache.spark.ml.regression中的类
-
AFTAggregator computes the gradient and loss for a AFT loss function,
as used in AFT survival regression for samples in sparse or dense vector in an online fashion.
- AFTAggregator(Broadcast<DenseVector<Object>>, boolean, Broadcast<double[]>) - 类 的构造器org.apache.spark.ml.regression.AFTAggregator
-
- AFTCostFun - org.apache.spark.ml.regression中的类
-
AFTCostFun implements Breeze's DiffFunction[T] for AFT cost.
- AFTCostFun(RDD<AFTPoint>, boolean, Broadcast<double[]>, int) - 类 的构造器org.apache.spark.ml.regression.AFTCostFun
-
- AFTSurvivalRegression - org.apache.spark.ml.regression中的类
-
- AFTSurvivalRegression(String) - 类 的构造器org.apache.spark.ml.regression.AFTSurvivalRegression
-
- AFTSurvivalRegression() - 类 的构造器org.apache.spark.ml.regression.AFTSurvivalRegression
-
- AFTSurvivalRegressionModel - org.apache.spark.ml.regression中的类
-
- agg(Column, Column...) - 类 中的方法org.apache.spark.sql.Dataset
-
Aggregates on the entire Dataset without groups.
- agg(Tuple2<String, String>, Seq<Tuple2<String, String>>) - 类 中的方法org.apache.spark.sql.Dataset
-
(Scala-specific) Aggregates on the entire Dataset without groups.
- agg(Map<String, String>) - 类 中的方法org.apache.spark.sql.Dataset
-
(Scala-specific) Aggregates on the entire Dataset without groups.
- agg(Map<String, String>) - 类 中的方法org.apache.spark.sql.Dataset
-
(Java-specific) Aggregates on the entire Dataset without groups.
- agg(Column, Seq<Column>) - 类 中的方法org.apache.spark.sql.Dataset
-
Aggregates on the entire Dataset without groups.
- agg(TypedColumn<V, U1>) - 类 中的方法org.apache.spark.sql.KeyValueGroupedDataset
-
Computes the given aggregation, returning a
Dataset of tuples for each unique key
and the result of computing this aggregation over all elements in the group.
- agg(TypedColumn<V, U1>, TypedColumn<V, U2>) - 类 中的方法org.apache.spark.sql.KeyValueGroupedDataset
-
Computes the given aggregations, returning a
Dataset of tuples for each unique key
and the result of computing these aggregations over all elements in the group.
- agg(TypedColumn<V, U1>, TypedColumn<V, U2>, TypedColumn<V, U3>) - 类 中的方法org.apache.spark.sql.KeyValueGroupedDataset
-
Computes the given aggregations, returning a
Dataset of tuples for each unique key
and the result of computing these aggregations over all elements in the group.
- agg(TypedColumn<V, U1>, TypedColumn<V, U2>, TypedColumn<V, U3>, TypedColumn<V, U4>) - 类 中的方法org.apache.spark.sql.KeyValueGroupedDataset
-
Computes the given aggregations, returning a
Dataset of tuples for each unique key
and the result of computing these aggregations over all elements in the group.
- agg(Column, Column...) - 类 中的方法org.apache.spark.sql.RelationalGroupedDataset
-
Compute aggregates by specifying a series of aggregate columns.
- agg(Tuple2<String, String>, Seq<Tuple2<String, String>>) - 类 中的方法org.apache.spark.sql.RelationalGroupedDataset
-
(Scala-specific) Compute aggregates by specifying the column names and
aggregate methods.
- agg(Map<String, String>) - 类 中的方法org.apache.spark.sql.RelationalGroupedDataset
-
(Scala-specific) Compute aggregates by specifying a map from column name to
aggregate methods.
- agg(Map<String, String>) - 类 中的方法org.apache.spark.sql.RelationalGroupedDataset
-
(Java-specific) Compute aggregates by specifying a map from column name to
aggregate methods.
- agg(Column, Seq<Column>) - 类 中的方法org.apache.spark.sql.RelationalGroupedDataset
-
Compute aggregates by specifying a series of aggregate columns.
- aggregate(U, Function2<U, T, U>, Function2<U, U, U>) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- aggregate(U, Function2<U, T, U>, Function2<U, U, U>) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- aggregate(U, Function2<U, T, U>, Function2<U, U, U>) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- aggregate(U, Function2<U, T, U>, Function2<U, U, U>) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Aggregate the elements of each partition, and then the results for all the partitions, using
given combine functions and a neutral "zero value".
- aggregate(U, Function2<U, T, U>, Function2<U, U, U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- aggregate(U, Function2<U, T, U>, Function2<U, U, U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- aggregate(U, Function2<U, T, U>, Function2<U, U, U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- aggregate(U, Function2<U, T, U>, Function2<U, U, U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- aggregate(U, Function2<U, T, U>, Function2<U, U, U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- aggregate(U, Function2<U, T, U>, Function2<U, U, U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- aggregate(U, Function2<U, T, U>, Function2<U, U, U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- aggregate(U, Function2<U, T, U>, Function2<U, U, U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- aggregate(U, Function2<U, T, U>, Function2<U, U, U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- aggregate(U, Function2<U, T, U>, Function2<U, U, U>, ClassTag<U>) - 类 中的方法org.apache.spark.rdd.RDD
-
Aggregate the elements of each partition, and then the results for all the partitions, using
given combine functions and a neutral "zero value".
- aggregate(U, Function2<U, T, U>, Function2<U, U, U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- aggregate(Function0<B>, Function2<B, A, B>, Function2<B, B, B>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- aggregateByKey(U, Partitioner, Function2<U, V, U>, Function2<U, U, U>) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Aggregate the values of each key, using given combine functions and a neutral "zero value".
- aggregateByKey(U, int, Function2<U, V, U>, Function2<U, U, U>) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Aggregate the values of each key, using given combine functions and a neutral "zero value".
- aggregateByKey(U, Function2<U, V, U>, Function2<U, U, U>) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Aggregate the values of each key, using given combine functions and a neutral "zero value".
- aggregateByKey(U, Partitioner, Function2<U, V, U>, Function2<U, U, U>, ClassTag<U>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Aggregate the values of each key, using given combine functions and a neutral "zero value".
- aggregateByKey(U, int, Function2<U, V, U>, Function2<U, U, U>, ClassTag<U>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Aggregate the values of each key, using given combine functions and a neutral "zero value".
- aggregateByKey(U, Function2<U, V, U>, Function2<U, U, U>, ClassTag<U>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Aggregate the values of each key, using given combine functions and a neutral "zero value".
- AggregatedDialect - org.apache.spark.sql.jdbc中的类
-
AggregatedDialect can unify multiple dialects into one virtual Dialect.
- AggregatedDialect(List<JdbcDialect>) - 类 的构造器org.apache.spark.sql.jdbc.AggregatedDialect
-
- aggregateMessages(Function1<EdgeContext<VD, ED, A>, BoxedUnit>, Function2<A, A, A>, TripletFields, ClassTag<A>) - 类 中的方法org.apache.spark.graphx.Graph
-
Aggregates values from the neighboring edges and vertices of each vertex.
- aggregateMessages(Function1<EdgeContext<VD, ED, A>, BoxedUnit>, Function2<A, A, A>, TripletFields, ClassTag<A>) - 类 中的静态方法org.apache.spark.graphx.impl.GraphImpl
-
- aggregateMessages$default$3() - 类 中的静态方法org.apache.spark.graphx.impl.GraphImpl
-
- aggregateMessagesWithActiveSet(Function1<EdgeContext<VD, ED, A>, BoxedUnit>, Function2<A, A, A>, TripletFields, Option<Tuple2<VertexRDD<?>, EdgeDirection>>, ClassTag<A>) - 类 中的方法org.apache.spark.graphx.impl.GraphImpl
-
- aggregateUsingIndex(RDD<Tuple2<Object, VD2>>, Function2<VD2, VD2, VD2>, ClassTag<VD2>) - 类 中的方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- aggregateUsingIndex(RDD<Tuple2<Object, VD2>>, Function2<VD2, VD2, VD2>, ClassTag<VD2>) - 类 中的方法org.apache.spark.graphx.VertexRDD
-
Aggregates vertices in messages that have the same ids using reduceFunc, returning a
VertexRDD co-indexed with this.
- AggregatingEdgeContext<VD,ED,A> - org.apache.spark.graphx.impl中的类
-
- AggregatingEdgeContext(Function2<A, A, A>, Object, BitSet) - 类 的构造器org.apache.spark.graphx.impl.AggregatingEdgeContext
-
- aggregationDepth() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- aggregationDepth() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- aggregationDepth() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- aggregationDepth() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- aggregationDepth() - 接口 中的方法org.apache.spark.ml.param.shared.HasAggregationDepth
-
Param for suggested depth for treeAggregate (>= 2).
- aggregationDepth() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- aggregationDepth() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- aggregationDepth() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- aggregationDepth() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- Aggregator<K,V,C> - org.apache.spark中的类
-
:: DeveloperApi ::
A set of functions used to aggregate data.
- Aggregator(Function1<V, C>, Function2<C, V, C>, Function2<C, C, C>) - 类 的构造器org.apache.spark.Aggregator
-
- aggregator() - 类 中的方法org.apache.spark.ShuffleDependency
-
- Aggregator<IN,BUF,OUT> - org.apache.spark.sql.expressions中的类
-
:: Experimental ::
A base class for user-defined aggregations, which can be used in Dataset operations to take
all of the elements of a group and reduce them to a single value.
- Aggregator() - 类 的构造器org.apache.spark.sql.expressions.Aggregator
-
- aic(RDD<Tuple3<Object, Object, Object>>, double, double, double) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.Binomial$
-
- aic(RDD<Tuple3<Object, Object, Object>>, double, double, double) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.Gamma$
-
- aic(RDD<Tuple3<Object, Object, Object>>, double, double, double) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.Gaussian$
-
- aic(RDD<Tuple3<Object, Object, Object>>, double, double, double) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.Poisson$
-
- aic() - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegressionSummary
-
Akaike Information Criterion (AIC) for the fitted model.
- Algo - org.apache.spark.mllib.tree.configuration中的类
-
Enum to select the algorithm for the decision tree
- Algo() - 类 的构造器org.apache.spark.mllib.tree.configuration.Algo
-
- algo() - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- algo() - 类 中的方法org.apache.spark.mllib.tree.model.DecisionTreeModel
-
- algo() - 类 中的方法org.apache.spark.mllib.tree.model.GradientBoostedTreesModel
-
- algo() - 类 中的方法org.apache.spark.mllib.tree.model.RandomForestModel
-
- algorithm() - 类 中的方法org.apache.spark.mllib.regression.StreamingLinearRegressionWithSGD
-
- alias(String) - 类 中的方法org.apache.spark.sql.Column
-
Gives the column an alias.
- alias(String) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new Dataset with an alias set.
- alias(Symbol) - 类 中的方法org.apache.spark.sql.Dataset
-
(Scala-specific) Returns a new Dataset with an alias set.
- aliases - 类 中的变量org.apache.spark.util.kvstore.LevelDB.TypeAliases
-
- All - 类 中的静态变量org.apache.spark.graphx.TripletFields
-
Expose all the fields (source, edge, and destination).
- allAttributes() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- allAttributes() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- allAttributes() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- allAttributes() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- AllJobsCancelled - org.apache.spark.scheduler中的类
-
- AllJobsCancelled() - 类 的构造器org.apache.spark.scheduler.AllJobsCancelled
-
- AllReceiverIds - org.apache.spark.streaming.scheduler中的类
-
A message used by ReceiverTracker to ask all receiver's ids still stored in
ReceiverTrackerEndpoint.
- AllReceiverIds() - 类 的构造器org.apache.spark.streaming.scheduler.AllReceiverIds
-
- allSources() - 类 中的静态方法org.apache.spark.metrics.source.StaticSources
-
The set of all static sources.
- alpha() - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- alpha() - 类 中的方法org.apache.spark.mllib.random.WeibullGenerator
-
- ALS - org.apache.spark.ml.recommendation中的类
-
Alternating Least Squares (ALS) matrix factorization.
- ALS(String) - 类 的构造器org.apache.spark.ml.recommendation.ALS
-
- ALS() - 类 的构造器org.apache.spark.ml.recommendation.ALS
-
- ALS - org.apache.spark.mllib.recommendation中的类
-
Alternating Least Squares matrix factorization.
- ALS() - 类 的构造器org.apache.spark.mllib.recommendation.ALS
-
Constructs an ALS instance with default parameters: {numBlocks: -1, rank: 10, iterations: 10,
lambda: 0.01, implicitPrefs: false, alpha: 1.0}.
- ALS.InBlock$ - org.apache.spark.ml.recommendation中的类
-
- ALS.Rating<ID> - org.apache.spark.ml.recommendation中的类
-
:: DeveloperApi ::
Rating class for better code readability.
- ALS.Rating$ - org.apache.spark.ml.recommendation中的类
-
- ALS.RatingBlock$ - org.apache.spark.ml.recommendation中的类
-
- ALSModel - org.apache.spark.ml.recommendation中的类
-
Model fitted by ALS.
- am() - 类 中的方法org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RegisterClusterManager
-
- AnalysisException - org.apache.spark.sql中的异常错误
-
Thrown when a query fails to analyze, usually because the query itself is invalid.
- and(Column) - 类 中的方法org.apache.spark.sql.Column
-
Boolean AND.
- And - org.apache.spark.sql.sources中的类
-
A filter that evaluates to true iff both left or right evaluate to true.
- And(Filter, Filter) - 类 的构造器org.apache.spark.sql.sources.And
-
- andThen(Function1<B, C>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- antecedent() - 类 中的方法org.apache.spark.mllib.fpm.AssociationRules.Rule
-
- ANY() - 类 中的静态方法org.apache.spark.scheduler.TaskLocality
-
- AnyDataType - org.apache.spark.sql.types中的类
-
An AbstractDataType that matches any concrete data types.
- AnyDataType() - 类 的构造器org.apache.spark.sql.types.AnyDataType
-
- anyNull() - 接口 中的方法org.apache.spark.sql.Row
-
Returns true if there are any NULL values in this row.
- anyNull() - 类 中的方法org.apache.spark.sql.vectorized.ColumnarRow
-
- ApiHelper - org.apache.spark.ui.jobs中的类
-
- ApiHelper() - 类 的构造器org.apache.spark.ui.jobs.ApiHelper
-
- appAttemptId() - 类 中的方法org.apache.spark.scheduler.SparkListenerApplicationStart
-
- Append() - 类 中的静态方法org.apache.spark.sql.streaming.OutputMode
-
OutputMode in which only the new rows in the streaming DataFrame/Dataset will be
written to the sink.
- appendBias(Vector) - 类 中的静态方法org.apache.spark.mllib.util.MLUtils
-
Returns a new vector with 1.0 (bias) appended to the input vector.
- appendColumn(StructType, String, DataType, boolean) - 类 中的静态方法org.apache.spark.ml.util.SchemaUtils
-
Appends a new column to the input schema.
- appendColumn(StructType, StructField) - 类 中的静态方法org.apache.spark.ml.util.SchemaUtils
-
Appends a new column to the input schema.
- appendReadColumns(Configuration, Seq<Integer>, Seq<String>) - 类 中的静态方法org.apache.spark.sql.hive.HiveShim
-
- appId() - 类 中的方法org.apache.spark.scheduler.SparkListenerApplicationStart
-
- APPLICATION_EXECUTOR_LIMIT() - 类 中的静态方法org.apache.spark.ui.ToolTips
-
- applicationAttemptId() - 类 中的方法org.apache.spark.SparkContext
-
- ApplicationAttemptInfo - org.apache.spark.status.api.v1中的类
-
- applicationEndFromJson(JsonAST.JValue) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- applicationEndToJson(SparkListenerApplicationEnd) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- ApplicationEnvironmentInfo - org.apache.spark.status.api.v1中的类
-
- applicationId() - 类 中的方法org.apache.spark.SparkContext
-
A unique identifier for the Spark application.
- ApplicationInfo - org.apache.spark.status.api.v1中的类
-
- applicationStartFromJson(JsonAST.JValue) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- applicationStartToJson(SparkListenerApplicationStart) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- ApplicationStatus - org.apache.spark.status.api.v1中的枚举
-
- apply(RDD<Tuple2<Object, VD>>, RDD<Edge<ED>>, VD, StorageLevel, StorageLevel, ClassTag<VD>, ClassTag<ED>) - 类 中的静态方法org.apache.spark.graphx.Graph
-
Construct a graph from a collection of vertices and
edges with attributes.
- apply(RDD<Edge<ED>>, VD, StorageLevel, StorageLevel, ClassTag<VD>, ClassTag<ED>) - 类 中的静态方法org.apache.spark.graphx.impl.GraphImpl
-
Create a graph from edges, setting referenced vertices to defaultVertexAttr.
- apply(RDD<Tuple2<Object, VD>>, RDD<Edge<ED>>, VD, StorageLevel, StorageLevel, ClassTag<VD>, ClassTag<ED>) - 类 中的静态方法org.apache.spark.graphx.impl.GraphImpl
-
Create a graph from vertices and edges, setting missing vertices to defaultVertexAttr.
- apply(VertexRDD<VD>, EdgeRDD<ED>, ClassTag<VD>, ClassTag<ED>) - 类 中的静态方法org.apache.spark.graphx.impl.GraphImpl
-
Create a graph from a VertexRDD and an EdgeRDD with arbitrary replicated vertices.
- apply(Graph<VD, ED>, A, int, EdgeDirection, Function3<Object, VD, A, VD>, Function1<EdgeTriplet<VD, ED>, Iterator<Tuple2<Object, A>>>, Function2<A, A, A>, ClassTag<VD>, ClassTag<ED>, ClassTag<A>) - 类 中的静态方法org.apache.spark.graphx.Pregel
-
Execute a Pregel-like iterative vertex-parallel abstraction.
- apply(RDD<Tuple2<Object, VD>>, ClassTag<VD>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
Constructs a standalone
VertexRDD (one that is not set up for efficient joins with an
EdgeRDD) from an RDD of vertex-attribute pairs.
- apply(RDD<Tuple2<Object, VD>>, EdgeRDD<?>, VD, ClassTag<VD>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
Constructs a VertexRDD from an RDD of vertex-attribute pairs.
- apply(RDD<Tuple2<Object, VD>>, EdgeRDD<?>, VD, Function2<VD, VD, VD>, ClassTag<VD>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
Constructs a VertexRDD from an RDD of vertex-attribute pairs.
- apply(DenseMatrix<Object>, DenseMatrix<Object>, Function1<Object, Object>) - 类 中的静态方法org.apache.spark.ml.ann.ApplyInPlace
-
- apply(DenseMatrix<Object>, DenseMatrix<Object>, DenseMatrix<Object>, Function2<Object, Object, Object>) - 类 中的静态方法org.apache.spark.ml.ann.ApplyInPlace
-
- apply(String) - 类 中的方法org.apache.spark.ml.attribute.AttributeGroup
-
Gets an attribute by its name.
- apply(int) - 类 中的方法org.apache.spark.ml.attribute.AttributeGroup
-
Gets an attribute by its index.
- apply(int, int) - 类 中的方法org.apache.spark.ml.linalg.DenseMatrix
-
- apply(int) - 类 中的方法org.apache.spark.ml.linalg.DenseVector
-
- apply(int, int) - 接口 中的方法org.apache.spark.ml.linalg.Matrix
-
Gets the (i, j)-th element.
- apply(int, int) - 类 中的方法org.apache.spark.ml.linalg.SparseMatrix
-
- apply(int) - 类 中的静态方法org.apache.spark.ml.linalg.SparseVector
-
- apply(int) - 接口 中的方法org.apache.spark.ml.linalg.Vector
-
Gets the value of the ith element.
- apply(Param<T>) - 类 中的方法org.apache.spark.ml.param.ParamMap
-
Gets the value of the input param or its default value if it does not exist.
- apply(GeneralizedLinearRegressionBase) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.FamilyAndLink$
-
Constructs the FamilyAndLink object from a parameter map
- apply(Split) - 类 中的方法org.apache.spark.ml.tree.DecisionTreeModelReadWrite.SplitData$
-
- apply(BinaryConfusionMatrix) - 类 中的静态方法org.apache.spark.mllib.evaluation.binary.FalsePositiveRate
-
- apply(BinaryConfusionMatrix) - 类 中的静态方法org.apache.spark.mllib.evaluation.binary.Precision
-
- apply(BinaryConfusionMatrix) - 类 中的静态方法org.apache.spark.mllib.evaluation.binary.Recall
-
- apply(int, int) - 类 中的方法org.apache.spark.mllib.linalg.DenseMatrix
-
- apply(int) - 类 中的方法org.apache.spark.mllib.linalg.DenseVector
-
- apply(int, int) - 接口 中的方法org.apache.spark.mllib.linalg.Matrix
-
Gets the (i, j)-th element.
- apply(int, int) - 类 中的方法org.apache.spark.mllib.linalg.SparseMatrix
-
- apply(int) - 类 中的静态方法org.apache.spark.mllib.linalg.SparseVector
-
- apply(int) - 接口 中的方法org.apache.spark.mllib.linalg.Vector
-
Gets the value of the ith element.
- apply(int) - 类 中的静态方法org.apache.spark.mllib.tree.configuration.Algo
-
- apply(int) - 类 中的静态方法org.apache.spark.mllib.tree.configuration.EnsembleCombiningStrategy
-
- apply(int) - 类 中的静态方法org.apache.spark.mllib.tree.configuration.FeatureType
-
- apply(int) - 类 中的静态方法org.apache.spark.mllib.tree.configuration.QuantileStrategy
-
- apply(int, Predict, double, boolean) - 类 中的静态方法org.apache.spark.mllib.tree.model.Node
-
Construct a node with nodeIndex, predict, impurity and isLeaf parameters.
- apply(int) - 类 中的静态方法org.apache.spark.rdd.CheckpointState
-
- apply(long, String, Option<String>, String, boolean) - 类 中的静态方法org.apache.spark.scheduler.AccumulableInfo
-
- apply(long, String, Option<String>, String) - 类 中的静态方法org.apache.spark.scheduler.AccumulableInfo
-
- apply(long, String, String) - 类 中的静态方法org.apache.spark.scheduler.AccumulableInfo
-
- apply(String, long, Enumeration.Value, ByteBuffer) - 类 中的方法org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.StatusUpdate$
-
Alternate factory method that takes a ByteBuffer directly for the data field
- apply(long, TaskMetrics) - 类 中的静态方法org.apache.spark.scheduler.RuntimePercentage
-
- apply(int) - 类 中的静态方法org.apache.spark.scheduler.SchedulingMode
-
- apply(int) - 类 中的静态方法org.apache.spark.scheduler.TaskLocality
-
- apply(Object) - 类 中的方法org.apache.spark.sql.Column
-
Extracts a value or values from a complex type.
- apply(String) - 类 中的方法org.apache.spark.sql.Dataset
-
Selects column based on the column name and returns it as a
Column.
- apply(Column...) - 类 中的方法org.apache.spark.sql.expressions.UserDefinedAggregateFunction
-
Creates a Column for this UDAF using given Columns as input arguments.
- apply(Seq<Column>) - 类 中的方法org.apache.spark.sql.expressions.UserDefinedAggregateFunction
-
Creates a Column for this UDAF using given Columns as input arguments.
- apply(Column...) - 类 中的方法org.apache.spark.sql.expressions.UserDefinedFunction
-
Returns an expression that invokes the UDF, using the given arguments.
- apply(Seq<Column>) - 类 中的方法org.apache.spark.sql.expressions.UserDefinedFunction
-
Returns an expression that invokes the UDF, using the given arguments.
- apply(LogicalPlan) - 类 中的方法org.apache.spark.sql.hive.DetermineTableStats
-
- apply(int) - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- apply(ScriptInputOutputSchema) - 类 中的静态方法org.apache.spark.sql.hive.execution.HiveScriptIOSchema
-
- apply(int) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- apply(int) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- apply(int) - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- apply(LogicalPlan) - 类 中的静态方法org.apache.spark.sql.hive.HiveAnalysis
-
- apply(LogicalPlan) - 类 中的方法org.apache.spark.sql.hive.RelationConversions
-
- apply(LogicalPlan) - 类 中的方法org.apache.spark.sql.hive.ResolveHiveSerdeTable
-
- apply(Dataset<Row>, Seq<Expression>, RelationalGroupedDataset.GroupType) - 类 中的静态方法org.apache.spark.sql.RelationalGroupedDataset
-
- apply(int) - 接口 中的方法org.apache.spark.sql.Row
-
Returns the value at position i.
- apply(String) - 类 中的静态方法org.apache.spark.sql.streaming.ProcessingTime
-
- apply(Duration) - 类 中的静态方法org.apache.spark.sql.streaming.ProcessingTime
-
- apply(DataType) - 类 中的静态方法org.apache.spark.sql.types.ArrayType
-
Construct a
ArrayType object with the given element type.
- apply(double) - 类 中的静态方法org.apache.spark.sql.types.Decimal
-
- apply(long) - 类 中的静态方法org.apache.spark.sql.types.Decimal
-
- apply(int) - 类 中的静态方法org.apache.spark.sql.types.Decimal
-
- apply(BigDecimal) - 类 中的静态方法org.apache.spark.sql.types.Decimal
-
- apply(BigDecimal) - 类 中的静态方法org.apache.spark.sql.types.Decimal
-
- apply(BigInteger) - 类 中的静态方法org.apache.spark.sql.types.Decimal
-
- apply(BigInt) - 类 中的静态方法org.apache.spark.sql.types.Decimal
-
- apply(BigDecimal, int, int) - 类 中的静态方法org.apache.spark.sql.types.Decimal
-
- apply(BigDecimal, int, int) - 类 中的静态方法org.apache.spark.sql.types.Decimal
-
- apply(long, int, int) - 类 中的静态方法org.apache.spark.sql.types.Decimal
-
- apply(String) - 类 中的静态方法org.apache.spark.sql.types.Decimal
-
- apply(DataType, DataType) - 类 中的静态方法org.apache.spark.sql.types.MapType
-
Construct a
MapType object with the given key type and value type.
- apply(String) - 类 中的方法org.apache.spark.sql.types.StructType
-
- apply(Set<String>) - 类 中的方法org.apache.spark.sql.types.StructType
-
Returns a
StructType containing
StructFields of the given names, preserving the
original order of fields.
- apply(int) - 类 中的方法org.apache.spark.sql.types.StructType
-
- apply(int) - 类 中的方法org.apache.spark.status.RDDPartitionSeq
-
- apply(String) - 类 中的静态方法org.apache.spark.storage.BlockId
-
- apply(String, String, int, Option<String>) - 类 中的静态方法org.apache.spark.storage.BlockManagerId
-
- apply(ObjectInput) - 类 中的静态方法org.apache.spark.storage.BlockManagerId
-
- apply(boolean, boolean, boolean, boolean, int) - 类 中的静态方法org.apache.spark.storage.StorageLevel
-
:: DeveloperApi ::
Create a new StorageLevel object.
- apply(boolean, boolean, boolean, int) - 类 中的静态方法org.apache.spark.storage.StorageLevel
-
:: DeveloperApi ::
Create a new StorageLevel object without setting useOffHeap.
- apply(int, int) - 类 中的静态方法org.apache.spark.storage.StorageLevel
-
:: DeveloperApi ::
Create a new StorageLevel object from its integer representation.
- apply(ObjectInput) - 类 中的静态方法org.apache.spark.storage.StorageLevel
-
:: DeveloperApi ::
Read StorageLevel object from ObjectInput stream.
- apply(long) - 类 中的静态方法org.apache.spark.streaming.Milliseconds
-
- apply(long) - 类 中的静态方法org.apache.spark.streaming.Minutes
-
- apply(int) - 类 中的静态方法org.apache.spark.streaming.scheduler.ReceiverState
-
- apply(long) - 类 中的静态方法org.apache.spark.streaming.Seconds
-
- apply(int) - 类 中的静态方法org.apache.spark.TaskState
-
- apply(TraversableOnce<Object>) - 类 中的静态方法org.apache.spark.util.StatCounter
-
Build a StatCounter from a list of values.
- apply(Seq<Object>) - 类 中的静态方法org.apache.spark.util.StatCounter
-
Build a StatCounter from a list of values passed as variable-length arguments.
- ApplyInPlace - org.apache.spark.ml.ann中的类
-
Implements in-place application of functions in the arrays
- ApplyInPlace() - 类 的构造器org.apache.spark.ml.ann.ApplyInPlace
-
- applyOrElse(A1, Function1<A1, B1>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- applySchema(RDD<Row>, StructType) - 类 中的方法org.apache.spark.sql.SQLContext
-
- applySchema(JavaRDD<Row>, StructType) - 类 中的方法org.apache.spark.sql.SQLContext
-
- applySchema(RDD<?>, Class<?>) - 类 中的方法org.apache.spark.sql.SQLContext
-
- applySchema(JavaRDD<?>, Class<?>) - 类 中的方法org.apache.spark.sql.SQLContext
-
- appName() - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
- appName() - 类 中的方法org.apache.spark.scheduler.SparkListenerApplicationStart
-
- appName() - 类 中的方法org.apache.spark.SparkContext
-
- appName(String) - 类 中的方法org.apache.spark.sql.SparkSession.Builder
-
Sets a name for the application, which will be shown in the Spark web UI.
- approx_count_distinct(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the approximate number of distinct items in a group.
- approx_count_distinct(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the approximate number of distinct items in a group.
- approx_count_distinct(Column, double) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the approximate number of distinct items in a group.
- approx_count_distinct(String, double) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the approximate number of distinct items in a group.
- approxCountDistinct(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
- approxCountDistinct(String) - 类 中的静态方法org.apache.spark.sql.functions
-
- approxCountDistinct(Column, double) - 类 中的静态方法org.apache.spark.sql.functions
-
- approxCountDistinct(String, double) - 类 中的静态方法org.apache.spark.sql.functions
-
- ApproxHist() - 类 中的静态方法org.apache.spark.mllib.tree.configuration.QuantileStrategy
-
- approxNearestNeighbors(Dataset<?>, Vector, int, String) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
-
- approxNearestNeighbors(Dataset<?>, Vector, int) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
-
- approxNearestNeighbors(Dataset<?>, Vector, int, String) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSHModel
-
- approxNearestNeighbors(Dataset<?>, Vector, int) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSHModel
-
- approxQuantile(String, double[], double) - 类 中的方法org.apache.spark.sql.DataFrameStatFunctions
-
Calculates the approximate quantiles of a numerical column of a DataFrame.
- approxQuantile(String[], double[], double) - 类 中的方法org.apache.spark.sql.DataFrameStatFunctions
-
Calculates the approximate quantiles of numerical columns of a DataFrame.
- approxSimilarityJoin(Dataset<?>, Dataset<?>, double, String) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
-
- approxSimilarityJoin(Dataset<?>, Dataset<?>, double) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
-
- approxSimilarityJoin(Dataset<?>, Dataset<?>, double, String) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSHModel
-
- approxSimilarityJoin(Dataset<?>, Dataset<?>, double) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSHModel
-
- appSparkVersion() - 类 中的方法org.apache.spark.status.api.v1.ApplicationAttemptInfo
-
- AppStatusUtils - org.apache.spark.status中的类
-
- AppStatusUtils() - 类 的构造器org.apache.spark.status.AppStatusUtils
-
- AreaUnderCurve - org.apache.spark.mllib.evaluation中的类
-
Computes the area under the curve (AUC) using the trapezoidal rule.
- AreaUnderCurve() - 类 的构造器org.apache.spark.mllib.evaluation.AreaUnderCurve
-
- areaUnderPR() - 类 中的方法org.apache.spark.mllib.evaluation.BinaryClassificationMetrics
-
Computes the area under the precision-recall curve.
- areaUnderROC() - 接口 中的方法org.apache.spark.ml.classification.BinaryLogisticRegressionSummary
-
Computes the area under the receiver operating characteristic (ROC) curve.
- areaUnderROC() - 类 中的方法org.apache.spark.mllib.evaluation.BinaryClassificationMetrics
-
Computes the area under the receiver operating characteristic (ROC) curve.
- argmax() - 类 中的方法org.apache.spark.ml.linalg.DenseVector
-
- argmax() - 类 中的方法org.apache.spark.ml.linalg.SparseVector
-
- argmax() - 接口 中的方法org.apache.spark.ml.linalg.Vector
-
Find the index of a maximal element.
- argmax() - 类 中的方法org.apache.spark.mllib.linalg.DenseVector
-
- argmax() - 类 中的方法org.apache.spark.mllib.linalg.SparseVector
-
- argmax() - 接口 中的方法org.apache.spark.mllib.linalg.Vector
-
Find the index of a maximal element.
- argString() - 类 中的方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- argString() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- argString() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- argString() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- array(DataType) - 类 中的方法org.apache.spark.sql.ColumnName
-
Creates a new StructField of type array.
- array(Column...) - 类 中的静态方法org.apache.spark.sql.functions
-
Creates a new array column.
- array(String, String...) - 类 中的静态方法org.apache.spark.sql.functions
-
Creates a new array column.
- array(Seq<Column>) - 类 中的静态方法org.apache.spark.sql.functions
-
Creates a new array column.
- array(String, Seq<String>) - 类 中的静态方法org.apache.spark.sql.functions
-
Creates a new array column.
- array() - 类 中的方法org.apache.spark.sql.vectorized.ColumnarArray
-
- array_contains(Column, Object) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns null if the array is null, true if the array contains value, and false otherwise.
- arrayLengthGt(double) - 类 中的静态方法org.apache.spark.ml.param.ParamValidators
-
Check that the array length is greater than lowerBound.
- ArrayType - org.apache.spark.sql.types中的类
-
- ArrayType(DataType, boolean) - 类 的构造器org.apache.spark.sql.types.ArrayType
-
- ArrowColumnVector - org.apache.spark.sql.vectorized中的类
-
A column vector backed by Apache Arrow.
- ArrowColumnVector(ValueVector) - 类 的构造器org.apache.spark.sql.vectorized.ArrowColumnVector
-
- as(Encoder<U>) - 类 中的方法org.apache.spark.sql.Column
-
Provides a type hint about the expected return value of this column.
- as(String) - 类 中的方法org.apache.spark.sql.Column
-
Gives the column an alias.
- as(Seq<String>) - 类 中的方法org.apache.spark.sql.Column
-
(Scala-specific) Assigns the given aliases to the results of a table generating function.
- as(String[]) - 类 中的方法org.apache.spark.sql.Column
-
Assigns the given aliases to the results of a table generating function.
- as(Symbol) - 类 中的方法org.apache.spark.sql.Column
-
Gives the column an alias.
- as(String, Metadata) - 类 中的方法org.apache.spark.sql.Column
-
Gives the column an alias with metadata.
- as(Encoder<U>) - 类 中的方法org.apache.spark.sql.Dataset
-
:: Experimental ::
Returns a new Dataset where each record has been mapped on to the specified type.
- as(String) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new Dataset with an alias set.
- as(Symbol) - 类 中的方法org.apache.spark.sql.Dataset
-
(Scala-specific) Returns a new Dataset with an alias set.
- asBinary() - 接口 中的方法org.apache.spark.ml.classification.LogisticRegressionSummary
-
Convenient method for casting to binary logistic regression summary.
- asBreeze() - 接口 中的方法org.apache.spark.ml.linalg.Matrix
-
Converts to a breeze matrix.
- asBreeze() - 接口 中的方法org.apache.spark.ml.linalg.Vector
-
Converts the instance to a breeze vector.
- asBreeze() - 接口 中的方法org.apache.spark.mllib.linalg.Matrix
-
Converts to a breeze matrix.
- asBreeze() - 接口 中的方法org.apache.spark.mllib.linalg.Vector
-
Converts the instance to a breeze vector.
- asc() - 类 中的方法org.apache.spark.sql.Column
-
Returns a sort expression based on ascending order of the column.
- asc(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns a sort expression based on ascending order of the column.
- asc_nulls_first() - 类 中的方法org.apache.spark.sql.Column
-
Returns a sort expression based on ascending order of the column,
and null values return before non-null values.
- asc_nulls_first(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns a sort expression based on ascending order of the column,
and null values return before non-null values.
- asc_nulls_last() - 类 中的方法org.apache.spark.sql.Column
-
Returns a sort expression based on ascending order of the column,
and null values appear after non-null values.
- asc_nulls_last(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns a sort expression based on ascending order of the column,
and null values appear after non-null values.
- ascii(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Computes the numeric value of the first character of the string column, and returns the
result as an int column.
- asCode() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- asCode() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- asCode() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- asCode() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- asin(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
- asin(String) - 类 中的静态方法org.apache.spark.sql.functions
-
- asIterator() - 类 中的方法org.apache.spark.serializer.DeserializationStream
-
Read the elements of this stream through an iterator.
- asJavaPairRDD() - 类 中的方法org.apache.spark.api.r.PairwiseRRDD
-
- asJavaRDD() - 类 中的方法org.apache.spark.api.r.RRDD
-
- asJavaRDD() - 类 中的方法org.apache.spark.api.r.StringRRDD
-
- asKeyValueIterator() - 类 中的方法org.apache.spark.serializer.DeserializationStream
-
Read the elements of this stream through an iterator over key-value pairs.
- AskPermissionToCommitOutput - org.apache.spark.scheduler中的类
-
- AskPermissionToCommitOutput(int, int, int) - 类 的构造器org.apache.spark.scheduler.AskPermissionToCommitOutput
-
- askRpcTimeout(SparkConf) - 类 中的静态方法org.apache.spark.util.RpcUtils
-
Returns the default Spark timeout to use for RPC ask operations.
- askSlaves() - 类 中的方法org.apache.spark.storage.BlockManagerMessages.GetBlockStatus
-
- askSlaves() - 类 中的方法org.apache.spark.storage.BlockManagerMessages.GetMatchingBlockIds
-
- asMap() - 类 中的方法org.apache.spark.sql.sources.v2.DataSourceOptions
-
- asML() - 类 中的方法org.apache.spark.mllib.linalg.DenseMatrix
-
- asML() - 类 中的方法org.apache.spark.mllib.linalg.DenseVector
-
- asML() - 接口 中的方法org.apache.spark.mllib.linalg.Matrix
-
Convert this matrix to the new mllib-local representation.
- asML() - 类 中的方法org.apache.spark.mllib.linalg.SparseMatrix
-
- asML() - 类 中的方法org.apache.spark.mllib.linalg.SparseVector
-
- asML() - 接口 中的方法org.apache.spark.mllib.linalg.Vector
-
Convert this vector to the new mllib-local representation.
- asNondeterministic() - 类 中的方法org.apache.spark.sql.expressions.UserDefinedFunction
-
Updates UserDefinedFunction to nondeterministic.
- asNonNullable() - 类 中的方法org.apache.spark.sql.expressions.UserDefinedFunction
-
Updates UserDefinedFunction to non-nullable.
- asNullable() - 类 中的方法org.apache.spark.sql.types.ObjectType
-
- asRDDId() - 类 中的方法org.apache.spark.storage.BlockId
-
- asRDDId() - 类 中的静态方法org.apache.spark.storage.BroadcastBlockId
-
- asRDDId() - 类 中的静态方法org.apache.spark.storage.RDDBlockId
-
- asRDDId() - 类 中的静态方法org.apache.spark.storage.ShuffleBlockId
-
- asRDDId() - 类 中的静态方法org.apache.spark.storage.ShuffleDataBlockId
-
- asRDDId() - 类 中的静态方法org.apache.spark.storage.ShuffleIndexBlockId
-
- asRDDId() - 类 中的静态方法org.apache.spark.storage.StreamBlockId
-
- asRDDId() - 类 中的静态方法org.apache.spark.storage.TaskResultBlockId
-
- assertConf(JobContext, SparkConf) - 类 中的方法org.apache.spark.internal.io.HadoopWriteConfigUtil
-
- assertNotSpilled(SparkContext, String, Function0<T>) - 类 中的静态方法org.apache.spark.TestUtils
-
Run some code involving jobs submitted to the given context and assert that the jobs
did not spill.
- assertSpilled(SparkContext, String, Function0<T>) - 类 中的静态方法org.apache.spark.TestUtils
-
Run some code involving jobs submitted to the given context and assert that the jobs spilled.
- Assignment(long, int) - 类 的构造器org.apache.spark.mllib.clustering.PowerIterationClustering.Assignment
-
- Assignment$() - 类 的构造器org.apache.spark.mllib.clustering.PowerIterationClustering.Assignment$
-
- assignments() - 类 中的方法org.apache.spark.mllib.clustering.PowerIterationClusteringModel
-
- AssociationRules - org.apache.spark.ml.fpm中的类
-
- AssociationRules() - 类 的构造器org.apache.spark.ml.fpm.AssociationRules
-
- associationRules() - 类 中的方法org.apache.spark.ml.fpm.FPGrowthModel
-
Get association rules fitted using the minConfidence.
- AssociationRules - org.apache.spark.mllib.fpm中的类
-
Generates association rules from a RDD[FreqItemset[Item}.
- AssociationRules() - 类 的构造器org.apache.spark.mllib.fpm.AssociationRules
-
Constructs a default instance with default parameters {minConfidence = 0.8}.
- AssociationRules.Rule<Item> - org.apache.spark.mllib.fpm中的类
-
An association rule between sets of items.
- ASYNC_TRACKING_ENABLED() - 类 中的静态方法org.apache.spark.status.config
-
- AsyncEventQueue - org.apache.spark.scheduler中的类
-
An asynchronous queue for events.
- AsyncEventQueue(String, SparkConf, LiveListenerBusMetrics, LiveListenerBus) - 类 的构造器org.apache.spark.scheduler.AsyncEventQueue
-
- AsyncRDDActions<T> - org.apache.spark.rdd中的类
-
A set of asynchronous RDD actions available through an implicit conversion.
- AsyncRDDActions(RDD<T>, ClassTag<T>) - 类 的构造器org.apache.spark.rdd.AsyncRDDActions
-
- atan(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
- atan(String) - 类 中的静态方法org.apache.spark.sql.functions
-
- atan2(Column, Column) - 类 中的静态方法org.apache.spark.sql.functions
-
- atan2(Column, String) - 类 中的静态方法org.apache.spark.sql.functions
-
- atan2(String, Column) - 类 中的静态方法org.apache.spark.sql.functions
-
- atan2(String, String) - 类 中的静态方法org.apache.spark.sql.functions
-
- atan2(Column, double) - 类 中的静态方法org.apache.spark.sql.functions
-
- atan2(String, double) - 类 中的静态方法org.apache.spark.sql.functions
-
- atan2(double, Column) - 类 中的静态方法org.apache.spark.sql.functions
-
- atan2(double, String) - 类 中的静态方法org.apache.spark.sql.functions
-
- attempt() - 类 中的方法org.apache.spark.status.api.v1.TaskData
-
- ATTEMPT() - 类 中的静态方法org.apache.spark.status.TaskIndexNames
-
- attemptId() - 类 中的方法org.apache.spark.scheduler.StageInfo
-
- attemptId() - 类 中的方法org.apache.spark.status.api.v1.ApplicationAttemptInfo
-
- attemptId() - 类 中的方法org.apache.spark.status.api.v1.StageData
-
- attemptNumber() - 类 中的方法org.apache.spark.scheduler.AskPermissionToCommitOutput
-
- attemptNumber() - 类 中的方法org.apache.spark.scheduler.StageInfo
-
- attemptNumber() - 类 中的方法org.apache.spark.scheduler.TaskInfo
-
- attemptNumber() - 类 中的方法org.apache.spark.TaskCommitDenied
-
- attemptNumber() - 类 中的方法org.apache.spark.TaskContext
-
How many times this task has been attempted.
- attempts() - 类 中的方法org.apache.spark.status.api.v1.ApplicationInfo
-
- AtTimestamp(Date) - 类 的构造器org.apache.spark.streaming.kinesis.KinesisInitialPositions.AtTimestamp
-
- attr() - 类 中的方法org.apache.spark.graphx.Edge
-
- attr() - 类 中的方法org.apache.spark.graphx.EdgeContext
-
The attribute associated with the edge.
- attr() - 类 中的方法org.apache.spark.graphx.impl.AggregatingEdgeContext
-
- Attribute - org.apache.spark.ml.attribute中的类
-
:: DeveloperApi ::
Abstract class for ML attributes.
- Attribute() - 类 的构造器org.apache.spark.ml.attribute.Attribute
-
- attribute() - 类 中的方法org.apache.spark.sql.sources.EqualNullSafe
-
- attribute() - 类 中的方法org.apache.spark.sql.sources.EqualTo
-
- attribute() - 类 中的方法org.apache.spark.sql.sources.GreaterThan
-
- attribute() - 类 中的方法org.apache.spark.sql.sources.GreaterThanOrEqual
-
- attribute() - 类 中的方法org.apache.spark.sql.sources.In
-
- attribute() - 类 中的方法org.apache.spark.sql.sources.IsNotNull
-
- attribute() - 类 中的方法org.apache.spark.sql.sources.IsNull
-
- attribute() - 类 中的方法org.apache.spark.sql.sources.LessThan
-
- attribute() - 类 中的方法org.apache.spark.sql.sources.LessThanOrEqual
-
- attribute() - 类 中的方法org.apache.spark.sql.sources.StringContains
-
- attribute() - 类 中的方法org.apache.spark.sql.sources.StringEndsWith
-
- attribute() - 类 中的方法org.apache.spark.sql.sources.StringStartsWith
-
- AttributeGroup - org.apache.spark.ml.attribute中的类
-
:: DeveloperApi ::
Attributes that describe a vector ML column.
- AttributeGroup(String) - 类 的构造器org.apache.spark.ml.attribute.AttributeGroup
-
Creates an attribute group without attribute info.
- AttributeGroup(String, int) - 类 的构造器org.apache.spark.ml.attribute.AttributeGroup
-
Creates an attribute group knowing only the number of attributes.
- AttributeGroup(String, Attribute[]) - 类 的构造器org.apache.spark.ml.attribute.AttributeGroup
-
Creates an attribute group with attributes.
- AttributeKeys - org.apache.spark.ml.attribute中的类
-
Keys used to store attributes.
- AttributeKeys() - 类 的构造器org.apache.spark.ml.attribute.AttributeKeys
-
- attributes() - 类 中的方法org.apache.spark.ml.attribute.AttributeGroup
-
Optional array of attributes.
- ATTRIBUTES() - 类 中的静态方法org.apache.spark.ml.attribute.AttributeKeys
-
- AttributeType - org.apache.spark.ml.attribute中的类
-
:: DeveloperApi ::
An enum-like type for attribute types: AttributeType$.Numeric, AttributeType$.Nominal,
and AttributeType$.Binary.
- AttributeType(String) - 类 的构造器org.apache.spark.ml.attribute.AttributeType
-
- attrType() - 类 中的方法org.apache.spark.ml.attribute.Attribute
-
Attribute type.
- attrType() - 类 中的方法org.apache.spark.ml.attribute.BinaryAttribute
-
- attrType() - 类 中的方法org.apache.spark.ml.attribute.NominalAttribute
-
- attrType() - 类 中的方法org.apache.spark.ml.attribute.NumericAttribute
-
- attrType() - 类 中的静态方法org.apache.spark.ml.attribute.UnresolvedAttribute
-
- available() - 类 中的方法org.apache.spark.io.NioBufferedFileInputStream
-
- available() - 类 中的方法org.apache.spark.io.ReadAheadInputStream
-
- available() - 类 中的方法org.apache.spark.storage.BufferReleasingInputStream
-
- Average() - 类 中的静态方法org.apache.spark.mllib.tree.configuration.EnsembleCombiningStrategy
-
- avg(MapFunction<T, Double>) - 类 中的静态方法org.apache.spark.sql.expressions.javalang.typed
-
Average aggregate function.
- avg(Function1<IN, Object>) - 类 中的静态方法org.apache.spark.sql.expressions.scalalang.typed
-
Average aggregate function.
- avg(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the average of the values in a group.
- avg(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the average of the values in a group.
- avg(String...) - 类 中的方法org.apache.spark.sql.RelationalGroupedDataset
-
Compute the mean value for each numeric columns for each group.
- avg(Seq<String>) - 类 中的方法org.apache.spark.sql.RelationalGroupedDataset
-
Compute the mean value for each numeric columns for each group.
- avg() - 类 中的方法org.apache.spark.util.DoubleAccumulator
-
Returns the average of elements added to the accumulator.
- avg() - 类 中的方法org.apache.spark.util.LongAccumulator
-
Returns the average of elements added to the accumulator.
- avgEventRate() - 类 中的方法org.apache.spark.status.api.v1.streaming.ReceiverInfo
-
- avgInputRate() - 类 中的方法org.apache.spark.status.api.v1.streaming.StreamingStatistics
-
- avgMetrics() - 类 中的方法org.apache.spark.ml.tuning.CrossValidatorModel
-
- avgProcessingTime() - 类 中的方法org.apache.spark.status.api.v1.streaming.StreamingStatistics
-
- avgSchedulingDelay() - 类 中的方法org.apache.spark.status.api.v1.streaming.StreamingStatistics
-
- avgTotalDelay() - 类 中的方法org.apache.spark.status.api.v1.streaming.StreamingStatistics
-
- awaitAnyTermination() - 类 中的方法org.apache.spark.sql.streaming.StreamingQueryManager
-
Wait until any of the queries on the associated SQLContext has terminated since the
creation of the context, or since resetTerminated() was called.
- awaitAnyTermination(long) - 类 中的方法org.apache.spark.sql.streaming.StreamingQueryManager
-
Wait until any of the queries on the associated SQLContext has terminated since the
creation of the context, or since resetTerminated() was called.
- awaitReady(Awaitable<T>, Duration) - 类 中的静态方法org.apache.spark.util.ThreadUtils
-
Preferred alternative to Await.ready().
- awaitResult(Awaitable<T>, Duration) - 类 中的静态方法org.apache.spark.util.ThreadUtils
-
Preferred alternative to Await.result().
- awaitTermination() - 接口 中的方法org.apache.spark.sql.streaming.StreamingQuery
-
Waits for the termination of this query, either by query.stop() or by an exception.
- awaitTermination(long) - 接口 中的方法org.apache.spark.sql.streaming.StreamingQuery
-
Waits for the termination of this query, either by query.stop() or by an exception.
- awaitTermination() - 类 中的方法org.apache.spark.streaming.api.java.JavaStreamingContext
-
Wait for the execution to stop.
- awaitTermination() - 类 中的方法org.apache.spark.streaming.StreamingContext
-
Wait for the execution to stop.
- awaitTerminationOrTimeout(long) - 类 中的方法org.apache.spark.streaming.api.java.JavaStreamingContext
-
Wait for the execution to stop.
- awaitTerminationOrTimeout(long) - 类 中的方法org.apache.spark.streaming.StreamingContext
-
Wait for the execution to stop.
- axpy(double, Vector, Vector) - 类 中的静态方法org.apache.spark.ml.linalg.BLAS
-
y += a * x
- axpy(double, Vector, Vector) - 类 中的静态方法org.apache.spark.mllib.linalg.BLAS
-
y += a * x
- cache() - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
Persist this RDD with the default storage level (MEMORY_ONLY).
- cache() - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Persist this RDD with the default storage level (MEMORY_ONLY).
- cache() - 类 中的方法org.apache.spark.api.java.JavaRDD
-
Persist this RDD with the default storage level (MEMORY_ONLY).
- cache() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- cache() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- cache() - 类 中的方法org.apache.spark.graphx.Graph
-
Caches the vertices and edges associated with this graph at the previously-specified target
storage levels, which default to MEMORY_ONLY.
- cache() - 类 中的方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
Persists the edge partitions using targetStorageLevel, which defaults to MEMORY_ONLY.
- cache() - 类 中的方法org.apache.spark.graphx.impl.GraphImpl
-
- cache() - 类 中的方法org.apache.spark.graphx.impl.VertexRDDImpl
-
Persists the vertex partitions at targetStorageLevel, which defaults to MEMORY_ONLY.
- cache() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- cache() - 类 中的方法org.apache.spark.mllib.linalg.distributed.BlockMatrix
-
Caches the underlying RDD.
- cache() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- cache() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- cache() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- cache() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- cache() - 类 中的方法org.apache.spark.rdd.RDD
-
Persist this RDD with the default storage level (MEMORY_ONLY).
- cache() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- cache() - 类 中的方法org.apache.spark.sql.Dataset
-
Persist this Dataset with the default storage level (MEMORY_AND_DISK).
- cache() - 类 中的方法org.apache.spark.streaming.api.java.JavaDStream
-
Persist RDDs of this DStream with the default storage level (MEMORY_ONLY_SER)
- cache() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- cache() - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Persist RDDs of this DStream with the default storage level (MEMORY_ONLY_SER)
- cache() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- cache() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- cache() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- cache() - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Persist RDDs of this DStream with the default storage level (MEMORY_ONLY_SER)
- cacheNodeIds() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- cacheNodeIds() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- cacheNodeIds() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- cacheNodeIds() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- cacheNodeIds() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- cacheNodeIds() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- cacheNodeIds() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- cacheNodeIds() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- cacheNodeIds() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- cacheNodeIds() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- cacheNodeIds() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- cacheNodeIds() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- cacheSize() - 接口 中的方法org.apache.spark.SparkExecutorInfo
-
- cacheSize() - 类 中的方法org.apache.spark.SparkExecutorInfoImpl
-
- cacheSize() - 类 中的方法org.apache.spark.storage.StorageStatus
-
已过时。
Return the memory used by caching RDDs
- cacheTable(String) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
Caches the specified table in-memory.
- cacheTable(String, StorageLevel) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
Caches the specified table with the given storage level.
- cacheTable(String) - 类 中的方法org.apache.spark.sql.SQLContext
-
Caches the specified table in-memory.
- calculate(DenseVector<Object>) - 类 中的方法org.apache.spark.ml.regression.AFTCostFun
-
- calculate(double[], double) - 类 中的静态方法org.apache.spark.mllib.tree.impurity.Entropy
-
:: DeveloperApi ::
information calculation for multiclass classification
- calculate(double, double, double) - 类 中的静态方法org.apache.spark.mllib.tree.impurity.Entropy
-
:: DeveloperApi ::
variance calculation
- calculate(double[], double) - 类 中的静态方法org.apache.spark.mllib.tree.impurity.Gini
-
:: DeveloperApi ::
information calculation for multiclass classification
- calculate(double, double, double) - 类 中的静态方法org.apache.spark.mllib.tree.impurity.Gini
-
:: DeveloperApi ::
variance calculation
- calculate(double[], double) - 接口 中的方法org.apache.spark.mllib.tree.impurity.Impurity
-
:: DeveloperApi ::
information calculation for multiclass classification
- calculate(double, double, double) - 接口 中的方法org.apache.spark.mllib.tree.impurity.Impurity
-
:: DeveloperApi ::
information calculation for regression
- calculate(double[], double) - 类 中的静态方法org.apache.spark.mllib.tree.impurity.Variance
-
:: DeveloperApi ::
information calculation for multiclass classification
- calculate(double, double, double) - 类 中的静态方法org.apache.spark.mllib.tree.impurity.Variance
-
:: DeveloperApi ::
variance calculation
- calculateNumberOfPartitions(long, int, int) - 类 中的方法org.apache.spark.ml.feature.Word2VecModel.Word2VecModelWriter$
-
Calculate the number of partitions to use in saving the model.
- CalendarIntervalType - org.apache.spark.sql.types中的类
-
The data type representing calendar time intervals.
- CalendarIntervalType - 类 中的静态变量org.apache.spark.sql.types.DataTypes
-
Gets the CalendarIntervalType object.
- call(K, Iterator<V1>, Iterator<V2>) - 接口 中的方法org.apache.spark.api.java.function.CoGroupFunction
-
- call(T) - 接口 中的方法org.apache.spark.api.java.function.DoubleFlatMapFunction
-
- call(T) - 接口 中的方法org.apache.spark.api.java.function.DoubleFunction
-
- call(T) - 接口 中的方法org.apache.spark.api.java.function.FilterFunction
-
- call(T) - 接口 中的方法org.apache.spark.api.java.function.FlatMapFunction
-
- call(T1, T2) - 接口 中的方法org.apache.spark.api.java.function.FlatMapFunction2
-
- call(K, Iterator<V>) - 接口 中的方法org.apache.spark.api.java.function.FlatMapGroupsFunction
-
- call(K, Iterator<V>, GroupState<S>) - 接口 中的方法org.apache.spark.api.java.function.FlatMapGroupsWithStateFunction
-
- call(T) - 接口 中的方法org.apache.spark.api.java.function.ForeachFunction
-
- call(Iterator<T>) - 接口 中的方法org.apache.spark.api.java.function.ForeachPartitionFunction
-
- call(T1) - 接口 中的方法org.apache.spark.api.java.function.Function
-
- call() - 接口 中的方法org.apache.spark.api.java.function.Function0
-
- call(T1, T2) - 接口 中的方法org.apache.spark.api.java.function.Function2
-
- call(T1, T2, T3) - 接口 中的方法org.apache.spark.api.java.function.Function3
-
- call(T1, T2, T3, T4) - 接口 中的方法org.apache.spark.api.java.function.Function4
-
- call(T) - 接口 中的方法org.apache.spark.api.java.function.MapFunction
-
- call(K, Iterator<V>) - 接口 中的方法org.apache.spark.api.java.function.MapGroupsFunction
-
- call(K, Iterator<V>, GroupState<S>) - 接口 中的方法org.apache.spark.api.java.function.MapGroupsWithStateFunction
-
- call(Iterator<T>) - 接口 中的方法org.apache.spark.api.java.function.MapPartitionsFunction
-
- call(T) - 接口 中的方法org.apache.spark.api.java.function.PairFlatMapFunction
-
- call(T) - 接口 中的方法org.apache.spark.api.java.function.PairFunction
-
- call(T, T) - 接口 中的方法org.apache.spark.api.java.function.ReduceFunction
-
- call(T) - 接口 中的方法org.apache.spark.api.java.function.VoidFunction
-
- call(T1, T2) - 接口 中的方法org.apache.spark.api.java.function.VoidFunction2
-
- call() - 接口 中的方法org.apache.spark.sql.api.java.UDF0
-
- call(T1) - 接口 中的方法org.apache.spark.sql.api.java.UDF1
-
- call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10) - 接口 中的方法org.apache.spark.sql.api.java.UDF10
-
- call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11) - 接口 中的方法org.apache.spark.sql.api.java.UDF11
-
- call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12) - 接口 中的方法org.apache.spark.sql.api.java.UDF12
-
- call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13) - 接口 中的方法org.apache.spark.sql.api.java.UDF13
-
- call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14) - 接口 中的方法org.apache.spark.sql.api.java.UDF14
-
- call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15) - 接口 中的方法org.apache.spark.sql.api.java.UDF15
-
- call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16) - 接口 中的方法org.apache.spark.sql.api.java.UDF16
-
- call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17) - 接口 中的方法org.apache.spark.sql.api.java.UDF17
-
- call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18) - 接口 中的方法org.apache.spark.sql.api.java.UDF18
-
- call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19) - 接口 中的方法org.apache.spark.sql.api.java.UDF19
-
- call(T1, T2) - 接口 中的方法org.apache.spark.sql.api.java.UDF2
-
- call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19, T20) - 接口 中的方法org.apache.spark.sql.api.java.UDF20
-
- call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19, T20, T21) - 接口 中的方法org.apache.spark.sql.api.java.UDF21
-
- call(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10, T11, T12, T13, T14, T15, T16, T17, T18, T19, T20, T21, T22) - 接口 中的方法org.apache.spark.sql.api.java.UDF22
-
- call(T1, T2, T3) - 接口 中的方法org.apache.spark.sql.api.java.UDF3
-
- call(T1, T2, T3, T4) - 接口 中的方法org.apache.spark.sql.api.java.UDF4
-
- call(T1, T2, T3, T4, T5) - 接口 中的方法org.apache.spark.sql.api.java.UDF5
-
- call(T1, T2, T3, T4, T5, T6) - 接口 中的方法org.apache.spark.sql.api.java.UDF6
-
- call(T1, T2, T3, T4, T5, T6, T7) - 接口 中的方法org.apache.spark.sql.api.java.UDF7
-
- call(T1, T2, T3, T4, T5, T6, T7, T8) - 接口 中的方法org.apache.spark.sql.api.java.UDF8
-
- call(T1, T2, T3, T4, T5, T6, T7, T8, T9) - 接口 中的方法org.apache.spark.sql.api.java.UDF9
-
- callSite() - 类 中的方法org.apache.spark.storage.RDDInfo
-
- callUDF(String, Column...) - 类 中的静态方法org.apache.spark.sql.functions
-
Call an user-defined function.
- callUDF(String, Seq<Column>) - 类 中的静态方法org.apache.spark.sql.functions
-
Call an user-defined function.
- cancel() - 类 中的方法org.apache.spark.ComplexFutureAction
-
- cancel() - 接口 中的方法org.apache.spark.FutureAction
-
Cancels the execution of this action.
- cancel() - 类 中的方法org.apache.spark.SimpleFutureAction
-
- cancelAllJobs() - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
Cancel all jobs that have been scheduled or are running.
- cancelAllJobs() - 类 中的方法org.apache.spark.SparkContext
-
Cancel all jobs that have been scheduled or are running.
- cancelJob(int, String) - 类 中的方法org.apache.spark.SparkContext
-
Cancel a given job if it's scheduled or running.
- cancelJob(int) - 类 中的方法org.apache.spark.SparkContext
-
Cancel a given job if it's scheduled or running.
- cancelJobGroup(String) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
Cancel active jobs for the specified group.
- cancelJobGroup(String) - 类 中的方法org.apache.spark.SparkContext
-
Cancel active jobs for the specified group.
- cancelStage(int, String) - 类 中的方法org.apache.spark.SparkContext
-
Cancel a given stage and all jobs associated with it.
- cancelStage(int) - 类 中的方法org.apache.spark.SparkContext
-
Cancel a given stage and all jobs associated with it.
- canEqual(Object) - 类 中的静态方法org.apache.spark.Aggregator
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.CleanAccum
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.CleanBroadcast
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.CleanCheckpoint
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.CleanRDD
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.CleanShuffle
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.ExceptionFailure
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.ExecutorLostFailure
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.ExecutorRegistered
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.ExecutorRemoved
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.ExpireDeadHosts
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.FetchFailed
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.graphx.Edge
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.ml.feature.Dot
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.ml.feature.LabeledPoint
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.ml.param.ParamPair
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.mllib.feature.VocabWord
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.mllib.linalg.distributed.IndexedRow
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.mllib.linalg.distributed.MatrixEntry
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.mllib.linalg.QRDecomposition
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.mllib.linalg.SingularValueDecomposition
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.mllib.recommendation.Rating
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.mllib.regression.LabeledPoint
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.mllib.stat.test.BinarySample
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.mllib.tree.configuration.BoostingStrategy
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.mllib.tree.model.Split
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.Resubmitted
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.rpc.netty.OnStart
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.rpc.netty.OnStop
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.AccumulableInfo
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.AllJobsCancelled
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.AskPermissionToCommitOutput
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.BlacklistedExecutor
-
- canEqual(Object) - 类 中的方法org.apache.spark.scheduler.cluster.ExecutorInfo
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.JobSucceeded
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.local.KillTask
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.local.ReviveOffers
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.local.StatusUpdate
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.local.StopExecutor
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.ResubmitFailedStages
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.RuntimePercentage
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerApplicationEnd
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerApplicationStart
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerBlockManagerAdded
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerBlockManagerRemoved
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerBlockUpdated
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerEnvironmentUpdate
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerExecutorAdded
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerExecutorBlacklisted
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerExecutorMetricsUpdate
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerExecutorRemoved
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerExecutorUnblacklisted
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerJobEnd
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerJobStart
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerLogStart
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerNodeBlacklisted
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerNodeUnblacklisted
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerSpeculativeTaskSubmitted
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerStageCompleted
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerStageSubmitted
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerTaskEnd
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerTaskGettingResult
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerTaskStart
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerUnpersistRDD
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.scheduler.StopCoordinator
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.DatasetHolder
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.expressions.UserDefinedFunction
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.hive.execution.HiveScriptIOSchema
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.hive.RelationConversions
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.jdbc.JdbcType
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.jdbc.MySQLDialect
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.jdbc.OracleDialect
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.jdbc.TeradataDialect
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.sources.And
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.sources.EqualNullSafe
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.sources.EqualTo
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.sources.GreaterThan
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.sources.GreaterThanOrEqual
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.sources.In
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.sources.IsNotNull
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.sources.IsNull
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.sources.LessThan
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.sources.LessThanOrEqual
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.sources.Not
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.sources.Or
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.sources.StringContains
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.sources.StringEndsWith
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.sources.StringStartsWith
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.streaming.ProcessingTime
-
已过时。
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.types.ArrayType
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.types.CharType
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.types.DecimalType
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.types.MapType
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.types.ObjectType
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.types.StructField
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.sql.types.VarcharType
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.status.api.v1.ApplicationAttemptInfo
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.status.api.v1.ApplicationInfo
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.StopMapOutputTracker
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.storage.BlockStatus
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.storage.BlockUpdatedInfo
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.storage.BroadcastBlockId
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.storage.memory.DeserializedMemoryEntry
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.storage.memory.SerializedMemoryEntry
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.storage.RDDBlockId
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.storage.ShuffleBlockId
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.storage.ShuffleDataBlockId
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.storage.ShuffleIndexBlockId
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.storage.StreamBlockId
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.storage.TaskResultBlockId
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.streaming.Duration
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.streaming.kinesis.DefaultCredentials
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.streaming.scheduler.AllReceiverIds
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.streaming.scheduler.BatchInfo
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.streaming.scheduler.GetAllReceiverInfo
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.streaming.scheduler.OutputOperationInfo
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.streaming.scheduler.ReceiverInfo
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.streaming.scheduler.StopAllReceivers
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerBatchCompleted
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerBatchStarted
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerBatchSubmitted
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerOutputOperationCompleted
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerOutputOperationStarted
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerReceiverError
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerReceiverStarted
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerReceiverStopped
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerStreamingStarted
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamInputInfo
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.streaming.Time
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.Success
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.TaskCommitDenied
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.TaskKilled
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.TaskResultLost
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.TaskSchedulerIsSet
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.UnknownReason
-
- canEqual(Object) - 类 中的静态方法org.apache.spark.util.MethodIdentifier
-
- canEqual(Object) - 类 中的方法org.apache.spark.util.MutablePair
-
- canHandle(String) - 类 中的方法org.apache.spark.sql.jdbc.AggregatedDialect
-
- canHandle(String) - 类 中的静态方法org.apache.spark.sql.jdbc.DB2Dialect
-
- canHandle(String) - 类 中的静态方法org.apache.spark.sql.jdbc.DerbyDialect
-
- canHandle(String) - 类 中的方法org.apache.spark.sql.jdbc.JdbcDialect
-
Check if this dialect instance can handle a certain jdbc url.
- canHandle(String) - 类 中的静态方法org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- canHandle(String) - 类 中的静态方法org.apache.spark.sql.jdbc.MySQLDialect
-
- canHandle(String) - 类 中的静态方法org.apache.spark.sql.jdbc.NoopDialect
-
- canHandle(String) - 类 中的静态方法org.apache.spark.sql.jdbc.OracleDialect
-
- canHandle(String) - 类 中的静态方法org.apache.spark.sql.jdbc.PostgresDialect
-
- canHandle(String) - 类 中的静态方法org.apache.spark.sql.jdbc.TeradataDialect
-
- canonicalized() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- canonicalized() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- canonicalized() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- canonicalized() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- CanonicalRandomVertexCut$() - 类 的构造器org.apache.spark.graphx.PartitionStrategy.CanonicalRandomVertexCut$
-
- cartesian(JavaRDDLike<U, ?>) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- cartesian(JavaRDDLike<U, ?>) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- cartesian(JavaRDDLike<U, ?>) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- cartesian(JavaRDDLike<U, ?>) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Return the Cartesian product of this RDD and another one, that is, the RDD of all pairs of
elements (a, b) where a is in this and b is in other.
- cartesian(RDD<U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- cartesian(RDD<U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- cartesian(RDD<U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- cartesian(RDD<U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- cartesian(RDD<U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- cartesian(RDD<U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- cartesian(RDD<U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- cartesian(RDD<U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- cartesian(RDD<U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- cartesian(RDD<U>, ClassTag<U>) - 类 中的方法org.apache.spark.rdd.RDD
-
Return the Cartesian product of this RDD and another one, that is, the RDD of all pairs of
elements (a, b) where a is in this and b is in other.
- cartesian(RDD<U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- caseSensitive() - 类 中的方法org.apache.spark.ml.feature.StopWordsRemover
-
Whether to do a case sensitive comparison over the stop words.
- cast(DataType) - 类 中的方法org.apache.spark.sql.Column
-
Casts the column to a different data type.
- cast(String) - 类 中的方法org.apache.spark.sql.Column
-
Casts the column to a different data type, using the canonical string representation
of the type.
- Catalog - org.apache.spark.sql.catalog中的类
-
Catalog interface for Spark.
- Catalog() - 类 的构造器org.apache.spark.sql.catalog.Catalog
-
- catalog() - 类 中的方法org.apache.spark.sql.SparkSession
-
Interface through which the user may create, drop, alter or query underlying
databases, tables, functions etc.
- catalogString() - 类 中的方法org.apache.spark.sql.types.ArrayType
-
- catalogString() - 类 中的静态方法org.apache.spark.sql.types.BinaryType
-
- catalogString() - 类 中的静态方法org.apache.spark.sql.types.BooleanType
-
- catalogString() - 类 中的静态方法org.apache.spark.sql.types.ByteType
-
- catalogString() - 类 中的静态方法org.apache.spark.sql.types.CalendarIntervalType
-
- catalogString() - 类 中的静态方法org.apache.spark.sql.types.CharType
-
- catalogString() - 类 中的方法org.apache.spark.sql.types.DataType
-
String representation for the type saved in external catalogs.
- catalogString() - 类 中的静态方法org.apache.spark.sql.types.DateType
-
- catalogString() - 类 中的静态方法org.apache.spark.sql.types.DecimalType
-
- catalogString() - 类 中的静态方法org.apache.spark.sql.types.DoubleType
-
- catalogString() - 类 中的静态方法org.apache.spark.sql.types.FloatType
-
- catalogString() - 类 中的静态方法org.apache.spark.sql.types.HiveStringType
-
- catalogString() - 类 中的静态方法org.apache.spark.sql.types.IntegerType
-
- catalogString() - 类 中的静态方法org.apache.spark.sql.types.LongType
-
- catalogString() - 类 中的方法org.apache.spark.sql.types.MapType
-
- catalogString() - 类 中的静态方法org.apache.spark.sql.types.NullType
-
- catalogString() - 类 中的静态方法org.apache.spark.sql.types.NumericType
-
- catalogString() - 类 中的静态方法org.apache.spark.sql.types.ObjectType
-
- catalogString() - 类 中的静态方法org.apache.spark.sql.types.ShortType
-
- catalogString() - 类 中的静态方法org.apache.spark.sql.types.StringType
-
- catalogString() - 类 中的方法org.apache.spark.sql.types.StructType
-
- catalogString() - 类 中的静态方法org.apache.spark.sql.types.TimestampType
-
- catalogString() - 类 中的静态方法org.apache.spark.sql.types.VarcharType
-
- CatalystScan - org.apache.spark.sql.sources中的接口
-
::Experimental::
An interface for experimenting with a more direct connection to the query planner.
- Categorical() - 类 中的静态方法org.apache.spark.mllib.tree.configuration.FeatureType
-
- categoricalCols() - 类 中的方法org.apache.spark.ml.feature.FeatureHasher
-
Numeric columns to treat as categorical features.
- categoricalFeaturesInfo() - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- CategoricalSplit - org.apache.spark.ml.tree中的类
-
Split which tests a categorical feature.
- categories() - 类 中的方法org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.SplitData
-
- categories() - 类 中的方法org.apache.spark.mllib.tree.model.Split
-
- categoryMaps() - 类 中的方法org.apache.spark.ml.feature.VectorIndexerModel
-
- categorySizes() - 类 中的方法org.apache.spark.ml.feature.OneHotEncoderModel
-
- cause() - 异常错误 中的方法org.apache.spark.sql.AnalysisException
-
- cause() - 异常错误 中的方法org.apache.spark.sql.streaming.StreamingQueryException
-
- CausedBy - org.apache.spark.util中的类
-
Extractor Object for pulling out the root cause of an error.
- CausedBy() - 类 的构造器org.apache.spark.util.CausedBy
-
- cbrt(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Computes the cube-root of the given value.
- cbrt(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Computes the cube-root of the given column.
- ceil(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Computes the ceiling of the given value.
- ceil(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Computes the ceiling of the given column.
- ceil() - 类 中的方法org.apache.spark.sql.types.Decimal
-
- censorCol() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- censorCol() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- chainl1(Function0<Parsers.Parser<T>>, Function0<Parsers.Parser<Function2<T, T, T>>>) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaParser
-
- chainl1(Function0<Parsers.Parser<T>>, Function0<Parsers.Parser<U>>, Function0<Parsers.Parser<Function2<T, U, T>>>) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaParser
-
- chainr1(Function0<Parsers.Parser<T>>, Function0<Parsers.Parser<Function2<T, U, U>>>, Function2<T, U, U>, U) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaParser
-
- changePrecision(int, int) - 类 中的方法org.apache.spark.sql.types.Decimal
-
Update precision and scale while keeping our value the same, and return true if successful.
- channelRead0(ChannelHandlerContext, byte[]) - 类 中的方法org.apache.spark.api.r.RBackendAuthHandler
-
- CharType - org.apache.spark.sql.types中的类
-
Hive char type.
- CharType(int) - 类 的构造器org.apache.spark.sql.types.CharType
-
- checkAndGetK8sMasterUrl(String) - 类 中的静态方法org.apache.spark.util.Utils
-
Check the validity of the given Kubernetes master URL and return the resolved URL.
- checkColumnNameDuplication(Seq<String>, String, Function2<String, String, Object>) - 类 中的静态方法org.apache.spark.sql.util.SchemaUtils
-
Checks if input column names have duplicate identifiers.
- checkColumnNameDuplication(Seq<String>, String, boolean) - 类 中的静态方法org.apache.spark.sql.util.SchemaUtils
-
Checks if input column names have duplicate identifiers.
- checkColumnType(StructType, String, DataType, String) - 类 中的静态方法org.apache.spark.ml.util.SchemaUtils
-
Check whether the given schema contains a column of the required data type.
- checkColumnTypes(StructType, String, Seq<DataType>, String) - 类 中的静态方法org.apache.spark.ml.util.SchemaUtils
-
Check whether the given schema contains a column of one of the require data types.
- checkDataColumns(RFormula, Dataset<?>) - 类 中的静态方法org.apache.spark.ml.r.RWrapperUtils
-
DataFrame column check.
- checkFileExists(String, Configuration) - 类 中的静态方法org.apache.spark.streaming.util.HdfsUtils
-
Check if the file exists at the given path.
- checkHost(String) - 类 中的静态方法org.apache.spark.util.Utils
-
- checkHostPort(String) - 类 中的静态方法org.apache.spark.util.Utils
-
- checkNumericType(StructType, String, String) - 类 中的静态方法org.apache.spark.ml.util.SchemaUtils
-
Check whether the given schema contains a column of the numeric data type.
- checkpoint() - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- checkpoint() - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- checkpoint() - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- checkpoint() - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Mark this RDD for checkpointing.
- checkpoint() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- checkpoint() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- checkpoint() - 类 中的方法org.apache.spark.graphx.Graph
-
Mark this Graph for checkpointing.
- checkpoint() - 类 中的方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- checkpoint() - 类 中的方法org.apache.spark.graphx.impl.GraphImpl
-
- checkpoint() - 类 中的方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- checkpoint() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- checkpoint() - 类 中的方法org.apache.spark.rdd.HadoopRDD
-
- checkpoint() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- checkpoint() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- checkpoint() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- checkpoint() - 类 中的方法org.apache.spark.rdd.RDD
-
Mark this RDD for checkpointing.
- checkpoint() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- checkpoint() - 类 中的方法org.apache.spark.sql.Dataset
-
Eagerly checkpoint a Dataset and return the new Dataset.
- checkpoint(boolean) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a checkpointed version of this Dataset.
- checkpoint(Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- checkpoint(Duration) - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
Enable periodic checkpointing of RDDs of this DStream.
- checkpoint(Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- checkpoint(Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- checkpoint(Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- checkpoint(Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- checkpoint(Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- checkpoint(String) - 类 中的方法org.apache.spark.streaming.api.java.JavaStreamingContext
-
Sets the context to periodically checkpoint the DStream operations for master
fault-tolerance.
- checkpoint(Duration) - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Enable periodic checkpointing of RDDs of this DStream
- checkpoint(String) - 类 中的方法org.apache.spark.streaming.StreamingContext
-
Set the context to periodically checkpoint the DStream operations for driver
fault-tolerance.
- Checkpointed() - 类 中的静态方法org.apache.spark.rdd.CheckpointState
-
- CheckpointingInProgress() - 类 中的静态方法org.apache.spark.rdd.CheckpointState
-
- checkpointInterval() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- checkpointInterval() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- checkpointInterval() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- checkpointInterval() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- checkpointInterval() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- checkpointInterval() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- checkpointInterval() - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- checkpointInterval() - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- checkpointInterval() - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- checkpointInterval() - 接口 中的方法org.apache.spark.ml.param.shared.HasCheckpointInterval
-
Param for set checkpoint interval (>= 1) or disable checkpoint (-1).
- checkpointInterval() - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- checkpointInterval() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- checkpointInterval() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- checkpointInterval() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- checkpointInterval() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- checkpointInterval() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- checkpointInterval() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- checkpointInterval() - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- CheckpointReader - org.apache.spark.streaming中的类
-
- CheckpointReader() - 类 的构造器org.apache.spark.streaming.CheckpointReader
-
- CheckpointState - org.apache.spark.rdd中的类
-
Enumeration to manage state transitions of an RDD through checkpointing
[ Initialized --> checkpointing in progress --> checkpointed ]
- CheckpointState() - 类 的构造器org.apache.spark.rdd.CheckpointState
-
- checkSchemaColumnNameDuplication(StructType, String, boolean) - 类 中的静态方法org.apache.spark.sql.util.SchemaUtils
-
Checks if an input schema has duplicate column names.
- checkSingleVsMultiColumnParams(Params, Seq<Param<?>>, Seq<Param<?>>) - 类 中的静态方法org.apache.spark.ml.param.ParamValidators
-
Utility for Param validity checks for Transformers which have both single- and multi-column
support.
- checkState(boolean, Function0<String>) - 类 中的静态方法org.apache.spark.streaming.util.HdfsUtils
-
- child() - 类 中的方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- child() - 类 中的方法org.apache.spark.sql.sources.Not
-
- CHILD_CONNECTION_TIMEOUT - 类 中的静态变量org.apache.spark.launcher.SparkLauncher
-
Maximum time (in ms) to wait for a child process to connect back to the launcher server
when using @link{#start()}.
- CHILD_PROCESS_LOGGER_NAME - 类 中的静态变量org.apache.spark.launcher.SparkLauncher
-
Logger name to use when launching a child process.
- children() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- children() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- children() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- children() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- childrenResolved() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- childrenResolved() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- childrenResolved() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- chiSqFunc() - 类 中的方法org.apache.spark.mllib.stat.test.ChiSqTest.Method
-
- ChiSqSelector - org.apache.spark.ml.feature中的类
-
Chi-Squared feature selection, which selects categorical features to use for predicting a
categorical label.
- ChiSqSelector(String) - 类 的构造器org.apache.spark.ml.feature.ChiSqSelector
-
- ChiSqSelector() - 类 的构造器org.apache.spark.ml.feature.ChiSqSelector
-
- ChiSqSelector - org.apache.spark.mllib.feature中的类
-
Creates a ChiSquared feature selector.
- ChiSqSelector() - 类 的构造器org.apache.spark.mllib.feature.ChiSqSelector
-
- ChiSqSelector(int) - 类 的构造器org.apache.spark.mllib.feature.ChiSqSelector
-
The is the same to call this() and setNumTopFeatures(numTopFeatures)
- ChiSqSelectorModel - org.apache.spark.ml.feature中的类
-
- ChiSqSelectorModel - org.apache.spark.mllib.feature中的类
-
Chi Squared selector model.
- ChiSqSelectorModel(int[]) - 类 的构造器org.apache.spark.mllib.feature.ChiSqSelectorModel
-
- ChiSqSelectorModel.SaveLoadV1_0$ - org.apache.spark.mllib.feature中的类
-
- ChiSqSelectorModel.SaveLoadV1_0$.Data - org.apache.spark.mllib.feature中的类
-
Model data for import/export
- chiSqTest(Vector, Vector) - 类 中的静态方法org.apache.spark.mllib.stat.Statistics
-
Conduct Pearson's chi-squared goodness of fit test of the observed data against the
expected distribution.
- chiSqTest(Vector) - 类 中的静态方法org.apache.spark.mllib.stat.Statistics
-
Conduct Pearson's chi-squared goodness of fit test of the observed data against the uniform
distribution, with each category having an expected frequency of 1 / observed.size.
- chiSqTest(Matrix) - 类 中的静态方法org.apache.spark.mllib.stat.Statistics
-
Conduct Pearson's independence test on the input contingency matrix, which cannot contain
negative entries or columns or rows that sum up to 0.
- chiSqTest(RDD<LabeledPoint>) - 类 中的静态方法org.apache.spark.mllib.stat.Statistics
-
Conduct Pearson's independence test for every feature against the label across the input RDD.
- chiSqTest(JavaRDD<LabeledPoint>) - 类 中的静态方法org.apache.spark.mllib.stat.Statistics
-
Java-friendly version of chiSqTest()
- ChiSqTest - org.apache.spark.mllib.stat.test中的类
-
Conduct the chi-squared test for the input RDDs using the specified method.
- ChiSqTest() - 类 的构造器org.apache.spark.mllib.stat.test.ChiSqTest
-
- ChiSqTest.Method - org.apache.spark.mllib.stat.test中的类
-
param: name String name for the method.
- ChiSqTest.Method$ - org.apache.spark.mllib.stat.test中的类
-
- ChiSqTest.NullHypothesis$ - org.apache.spark.mllib.stat.test中的类
-
- ChiSqTestResult - org.apache.spark.mllib.stat.test中的类
-
Object containing the test results for the chi-squared hypothesis test.
- chiSquared(Vector, Vector, String) - 类 中的静态方法org.apache.spark.mllib.stat.test.ChiSqTest
-
- chiSquaredFeatures(RDD<LabeledPoint>, String) - 类 中的静态方法org.apache.spark.mllib.stat.test.ChiSqTest
-
Conduct Pearson's independence test for each feature against the label across the input RDD.
- chiSquaredMatrix(Matrix, String) - 类 中的静态方法org.apache.spark.mllib.stat.test.ChiSqTest
-
- ChiSquareTest - org.apache.spark.ml.stat中的类
-
:: Experimental ::
Chi-square hypothesis testing for categorical data.
- ChiSquareTest() - 类 的构造器org.apache.spark.ml.stat.ChiSquareTest
-
- chmod700(File) - 类 中的静态方法org.apache.spark.util.Utils
-
JDK equivalent of chmod 700 file.
- CholeskyDecomposition - org.apache.spark.mllib.linalg中的类
-
Compute Cholesky decomposition.
- CholeskyDecomposition() - 类 的构造器org.apache.spark.mllib.linalg.CholeskyDecomposition
-
- classForName(String) - 类 中的静态方法org.apache.spark.util.Utils
-
Preferred alternative to Class.forName(className)
- Classification() - 类 中的静态方法org.apache.spark.mllib.tree.configuration.Algo
-
- ClassificationModel<FeaturesType,M extends ClassificationModel<FeaturesType,M>> - org.apache.spark.ml.classification中的类
-
:: DeveloperApi ::
Model produced by a
Classifier.
- ClassificationModel() - 类 的构造器org.apache.spark.ml.classification.ClassificationModel
-
- ClassificationModel - org.apache.spark.mllib.classification中的接口
-
Represents a classification model that predicts to which of a set of categories an example
belongs.
- Classifier<FeaturesType,E extends Classifier<FeaturesType,E,M>,M extends ClassificationModel<FeaturesType,M>> - org.apache.spark.ml.classification中的类
-
:: DeveloperApi ::
Single-label binary or multiclass classification.
- Classifier() - 类 的构造器org.apache.spark.ml.classification.Classifier
-
- classifier() - 类 中的静态方法org.apache.spark.ml.classification.OneVsRest
-
- classifier() - 类 中的静态方法org.apache.spark.ml.classification.OneVsRestModel
-
- classIsLoadable(String) - 类 中的静态方法org.apache.spark.util.Utils
-
Determines whether the provided class is loadable in the current thread.
- className() - 类 中的方法org.apache.spark.ExceptionFailure
-
- className() - 类 中的静态方法org.apache.spark.ml.linalg.JsonMatrixConverter
-
Unique class name for identifying JSON object encoded by this class.
- className() - 类 中的方法org.apache.spark.sql.catalog.Function
-
- classpathEntries() - 类 中的方法org.apache.spark.status.api.v1.ApplicationEnvironmentInfo
-
- classTag() - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
- classTag() - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
- classTag() - 类 中的方法org.apache.spark.api.java.JavaRDD
-
- classTag() - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
- classTag() - 类 中的方法org.apache.spark.sql.Dataset
-
- classTag() - 类 中的方法org.apache.spark.storage.memory.DeserializedMemoryEntry
-
- classTag() - 接口 中的方法org.apache.spark.storage.memory.MemoryEntry
-
- classTag() - 类 中的方法org.apache.spark.storage.memory.SerializedMemoryEntry
-
- classTag() - 类 中的方法org.apache.spark.streaming.api.java.JavaDStream
-
- classTag() - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
- classTag() - 类 中的方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- classTag() - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- classTag() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- classTag() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- classTag() - 类 中的方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- clean(long, boolean) - 类 中的方法org.apache.spark.streaming.util.WriteAheadLog
-
Clean all the records that are older than the threshold time.
- clean(Object, boolean, boolean) - 类 中的静态方法org.apache.spark.util.ClosureCleaner
-
Clean the given closure in place.
- CleanAccum - org.apache.spark中的类
-
- CleanAccum(long) - 类 的构造器org.apache.spark.CleanAccum
-
- CleanBroadcast - org.apache.spark中的类
-
- CleanBroadcast(long) - 类 的构造器org.apache.spark.CleanBroadcast
-
- CleanCheckpoint - org.apache.spark中的类
-
- CleanCheckpoint(int) - 类 的构造器org.apache.spark.CleanCheckpoint
-
- cleaning() - 类 中的方法org.apache.spark.status.LiveStage
-
- CleanRDD - org.apache.spark中的类
-
- CleanRDD(int) - 类 的构造器org.apache.spark.CleanRDD
-
- CleanShuffle - org.apache.spark中的类
-
- CleanShuffle(int) - 类 的构造器org.apache.spark.CleanShuffle
-
- CleanupTask - org.apache.spark中的接口
-
Classes that represent cleaning tasks.
- CleanupTaskWeakReference - org.apache.spark中的类
-
A WeakReference associated with a CleanupTask.
- CleanupTaskWeakReference(CleanupTask, Object, ReferenceQueue<Object>) - 类 的构造器org.apache.spark.CleanupTaskWeakReference
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.classification.OneVsRest
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.classification.OneVsRestModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeans
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixture
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.clustering.KMeans
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.clustering.KMeansModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.evaluation.ClusteringEvaluator
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.evaluation.RegressionEvaluator
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.Binarizer
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.Bucketizer
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelector
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.ColumnPruner
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizer
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizerModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.DCT
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.ElementwiseProduct
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.FeatureHasher
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.HashingTF
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.IDF
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.IDFModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.Imputer
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.ImputerModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.IndexToString
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.Interaction
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.MaxAbsScaler
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.MaxAbsScalerModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSH
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSHModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScaler
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScalerModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.NGram
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.Normalizer
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoder
-
已过时。
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderEstimator
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.PCA
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.PCAModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.PolynomialExpansion
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.QuantileDiscretizer
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.RegexTokenizer
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.RFormula
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.SQLTransformer
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.StandardScaler
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.StandardScalerModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.StopWordsRemover
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.StringIndexer
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.StringIndexerModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.Tokenizer
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.VectorAssembler
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.VectorAttributeRewriter
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexer
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexerModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.VectorSizeHint
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.VectorSlicer
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.Word2Vec
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.feature.Word2VecModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowth
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowthModel
-
- clear(Param<?>) - 接口 中的方法org.apache.spark.ml.param.Params
-
Clears the user-supplied value for the input param.
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.Pipeline
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.PipelineModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.recommendation.ALSModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegression
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegressionModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidator
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidatorModel
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplit
-
- clear(Param<?>) - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplitModel
-
- clear() - 类 中的方法org.apache.spark.sql.util.ExecutionListenerManager
-
- clear() - 类 中的静态方法org.apache.spark.util.AccumulatorContext
-
- clearActive() - 类 中的静态方法org.apache.spark.sql.SQLContext
-
- clearActiveSession() - 类 中的静态方法org.apache.spark.sql.SparkSession
-
Clears the active SparkSession for current thread.
- clearCache() - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
Removes all cached tables from the in-memory cache.
- clearCache() - 类 中的方法org.apache.spark.sql.SQLContext
-
Removes all cached tables from the in-memory cache.
- clearCallSite() - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
Pass-through to SparkContext.setCallSite.
- clearCallSite() - 类 中的方法org.apache.spark.SparkContext
-
Clear the thread-local property for overriding the call sites
of actions and RDDs.
- clearDefaultSession() - 类 中的静态方法org.apache.spark.sql.SparkSession
-
Clears the default SparkSession that is returned by the builder.
- clearDependencies() - 类 中的方法org.apache.spark.rdd.CoGroupedRDD
-
- clearDependencies() - 类 中的方法org.apache.spark.rdd.ShuffledRDD
-
- clearDependencies() - 类 中的方法org.apache.spark.rdd.UnionRDD
-
- clearJobGroup() - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
Clear the current thread's job group ID and its description.
- clearJobGroup() - 类 中的方法org.apache.spark.SparkContext
-
Clear the current thread's job group ID and its description.
- clearThreshold() - 类 中的方法org.apache.spark.mllib.classification.LogisticRegressionModel
-
Clears the threshold so that predict will output raw prediction scores.
- clearThreshold() - 类 中的方法org.apache.spark.mllib.classification.SVMModel
-
Clears the threshold so that predict will output raw prediction scores.
- CLogLog$() - 类 的构造器org.apache.spark.ml.regression.GeneralizedLinearRegression.CLogLog$
-
- clone() - 类 中的方法org.apache.spark.SparkConf
-
Copy this object
- clone() - 类 中的方法org.apache.spark.sql.ExperimentalMethods
-
- clone() - 类 中的方法org.apache.spark.sql.types.Decimal
-
- clone() - 类 中的方法org.apache.spark.sql.util.ExecutionListenerManager
-
Get an identical copy of this listener manager.
- clone() - 类 中的方法org.apache.spark.storage.StorageLevel
-
- clone() - 类 中的方法org.apache.spark.util.random.BernoulliCellSampler
-
- clone() - 类 中的方法org.apache.spark.util.random.BernoulliSampler
-
- clone() - 类 中的方法org.apache.spark.util.random.PoissonSampler
-
- clone() - 接口 中的方法org.apache.spark.util.random.RandomSampler
-
return a copy of the RandomSampler object
- clone(T, SerializerInstance, ClassTag<T>) - 类 中的静态方法org.apache.spark.util.Utils
-
Clone an object using a Spark serializer.
- cloneComplement() - 类 中的方法org.apache.spark.util.random.BernoulliCellSampler
-
Return a sampler that is the complement of the range specified of the current sampler.
- close() - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
- close() - 类 中的方法org.apache.spark.io.NioBufferedFileInputStream
-
- close() - 类 中的方法org.apache.spark.io.ReadAheadInputStream
-
- close() - 类 中的方法org.apache.spark.io.SnappyOutputStreamWrapper
-
- close() - 类 中的方法org.apache.spark.serializer.DeserializationStream
-
- close() - 类 中的方法org.apache.spark.serializer.SerializationStream
-
- close(Throwable) - 类 中的方法org.apache.spark.sql.ForeachWriter
-
Called when stopping to process one partition of new data in the executor side.
- close() - 类 中的方法org.apache.spark.sql.hive.execution.HiveOutputWriter
-
- close() - 类 中的方法org.apache.spark.sql.SparkSession
-
Synonym for stop().
- close() - 类 中的方法org.apache.spark.sql.vectorized.ArrowColumnVector
-
- close() - 类 中的方法org.apache.spark.sql.vectorized.ColumnarBatch
-
Called to close all the columns in this batch.
- close() - 类 中的方法org.apache.spark.sql.vectorized.ColumnVector
-
Cleans up memory for this column vector.
- close() - 类 中的方法org.apache.spark.storage.BufferReleasingInputStream
-
- close() - 类 中的方法org.apache.spark.storage.CountingWritableChannel
-
- close() - 类 中的方法org.apache.spark.storage.TimeTrackingOutputStream
-
- close() - 类 中的方法org.apache.spark.streaming.api.java.JavaStreamingContext
-
- close() - 类 中的方法org.apache.spark.streaming.util.WriteAheadLog
-
Close this log and release any resources.
- close() - 类 中的方法org.apache.spark.util.kvstore.InMemoryStore
-
- close() - 类 中的方法org.apache.spark.util.kvstore.LevelDB
-
- closeableIterator() - 类 中的方法org.apache.spark.util.kvstore.KVStoreView
-
Returns an iterator for the current configuration.
- closeWriter(TaskAttemptContext) - 类 中的方法org.apache.spark.internal.io.HadoopWriteConfigUtil
-
- ClosureCleaner - org.apache.spark.util中的类
-
A cleaner that renders closures serializable if they can be done so safely.
- ClosureCleaner() - 类 的构造器org.apache.spark.util.ClosureCleaner
-
- closureSerializer() - 类 中的方法org.apache.spark.SparkEnv
-
- cls() - 类 中的方法org.apache.spark.sql.types.ObjectType
-
- cls() - 类 中的方法org.apache.spark.util.MethodIdentifier
-
- clsTag() - 接口 中的方法org.apache.spark.sql.Encoder
-
A ClassTag that can be used to construct an Array to contain a collection of T.
- cluster() - 类 中的方法org.apache.spark.ml.clustering.ClusteringSummary
-
Cluster centers of the transformed data.
- cluster() - 类 中的方法org.apache.spark.mllib.clustering.PowerIterationClustering.Assignment
-
- clusterCenters() - 类 中的方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
- clusterCenters() - 类 中的方法org.apache.spark.ml.clustering.KMeansModel
-
- clusterCenters() - 类 中的方法org.apache.spark.mllib.clustering.BisectingKMeansModel
-
Leaf cluster centers.
- clusterCenters() - 类 中的方法org.apache.spark.mllib.clustering.KMeansModel
-
- clusterCenters() - 类 中的方法org.apache.spark.mllib.clustering.StreamingKMeansModel
-
- clusteredColumns - 类 中的变量org.apache.spark.sql.sources.v2.reader.partitioning.ClusteredDistribution
-
The names of the clustered columns.
- ClusteredDistribution - org.apache.spark.sql.sources.v2.reader.partitioning中的类
-
- ClusteredDistribution(String[]) - 类 的构造器org.apache.spark.sql.sources.v2.reader.partitioning.ClusteredDistribution
-
- ClusteringEvaluator - org.apache.spark.ml.evaluation中的类
-
:: Experimental ::
Evaluator for clustering results.
- ClusteringEvaluator(String) - 类 的构造器org.apache.spark.ml.evaluation.ClusteringEvaluator
-
- ClusteringEvaluator() - 类 的构造器org.apache.spark.ml.evaluation.ClusteringEvaluator
-
- ClusteringSummary - org.apache.spark.ml.clustering中的类
-
:: Experimental ::
Summary of clustering algorithms.
- clusterSizes() - 类 中的方法org.apache.spark.ml.clustering.ClusteringSummary
-
Size of (number of data points in) each cluster.
- ClusterStats(Vector, double, long) - 类 的构造器org.apache.spark.ml.evaluation.SquaredEuclideanSilhouette.ClusterStats
-
- ClusterStats$() - 类 的构造器org.apache.spark.ml.evaluation.SquaredEuclideanSilhouette.ClusterStats$
-
- clusterWeights() - 类 中的方法org.apache.spark.mllib.clustering.StreamingKMeansModel
-
- cn() - 类 中的方法org.apache.spark.mllib.feature.VocabWord
-
- coalesce(int) - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
Return a new RDD that is reduced into numPartitions partitions.
- coalesce(int, boolean) - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
Return a new RDD that is reduced into numPartitions partitions.
- coalesce(int) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Return a new RDD that is reduced into numPartitions partitions.
- coalesce(int, boolean) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Return a new RDD that is reduced into numPartitions partitions.
- coalesce(int) - 类 中的方法org.apache.spark.api.java.JavaRDD
-
Return a new RDD that is reduced into numPartitions partitions.
- coalesce(int, boolean) - 类 中的方法org.apache.spark.api.java.JavaRDD
-
Return a new RDD that is reduced into numPartitions partitions.
- coalesce(int, boolean, Option<PartitionCoalescer>, Ordering<T>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- coalesce(int, boolean, Option<PartitionCoalescer>, Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- coalesce(int, boolean, Option<PartitionCoalescer>, Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- coalesce(int, boolean, Option<PartitionCoalescer>, Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- coalesce(int, boolean, Option<PartitionCoalescer>, Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- coalesce(int, RDD<?>) - 类 中的方法org.apache.spark.rdd.DefaultPartitionCoalescer
-
Runs the packing algorithm and returns an array of PartitionGroups that if possible are
load balanced and grouped by locality
- coalesce(int, boolean, Option<PartitionCoalescer>, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- coalesce(int, boolean, Option<PartitionCoalescer>, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- coalesce(int, boolean, Option<PartitionCoalescer>, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- coalesce(int, RDD<?>) - 接口 中的方法org.apache.spark.rdd.PartitionCoalescer
-
Coalesce the partitions of the given RDD.
- coalesce(int, boolean, Option<PartitionCoalescer>, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- coalesce(int, boolean, Option<PartitionCoalescer>, Ordering<T>) - 类 中的方法org.apache.spark.rdd.RDD
-
Return a new RDD that is reduced into numPartitions partitions.
- coalesce(int, boolean, Option<PartitionCoalescer>, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- coalesce(int) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new Dataset that has exactly numPartitions partitions, when the fewer partitions
are requested.
- coalesce(Column...) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the first column that is not null, or null if all inputs are null.
- coalesce(Seq<Column>) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the first column that is not null, or null if all inputs are null.
- coalesce$default$2() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- coalesce$default$2() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- coalesce$default$2() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- coalesce$default$2() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- coalesce$default$2() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- coalesce$default$2() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- coalesce$default$2() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- coalesce$default$2() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- coalesce$default$2() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- coalesce$default$2() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- coalesce$default$3() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- coalesce$default$3() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- coalesce$default$3() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- coalesce$default$3() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- coalesce$default$3() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- coalesce$default$3() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- coalesce$default$3() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- coalesce$default$3() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- coalesce$default$3() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- coalesce$default$3() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- coalesce$default$4(int, boolean, Option<PartitionCoalescer>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- coalesce$default$4(int, boolean, Option<PartitionCoalescer>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- coalesce$default$4(int, boolean, Option<PartitionCoalescer>) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- coalesce$default$4(int, boolean, Option<PartitionCoalescer>) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- coalesce$default$4(int, boolean, Option<PartitionCoalescer>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- coalesce$default$4(int, boolean, Option<PartitionCoalescer>) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- coalesce$default$4(int, boolean, Option<PartitionCoalescer>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- coalesce$default$4(int, boolean, Option<PartitionCoalescer>) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- coalesce$default$4(int, boolean, Option<PartitionCoalescer>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- coalesce$default$4(int, boolean, Option<PartitionCoalescer>) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- CoarseGrainedClusterMessages - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages() - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages
-
- CoarseGrainedClusterMessages.AddWebUIFilter - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.AddWebUIFilter$ - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.GetExecutorLossReason - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.GetExecutorLossReason$ - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.KillExecutors - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.KillExecutors$ - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.KillExecutorsOnHost - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.KillExecutorsOnHost$ - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.KillTask - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.KillTask$ - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.LaunchTask - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.LaunchTask$ - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.RegisterClusterManager - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.RegisterClusterManager$ - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.RegisteredExecutor$ - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.RegisterExecutor - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.RegisterExecutor$ - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.RegisterExecutorFailed - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.RegisterExecutorFailed$ - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.RegisterExecutorResponse - org.apache.spark.scheduler.cluster中的接口
-
- CoarseGrainedClusterMessages.RemoveExecutor - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.RemoveExecutor$ - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.RemoveWorker - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.RemoveWorker$ - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.RequestExecutors - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.RequestExecutors$ - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.RetrieveLastAllocatedExecutorId$ - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.RetrieveSparkAppConfig$ - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.ReviveOffers$ - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.SetupDriver - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.SetupDriver$ - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.Shutdown$ - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.SparkAppConfig - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.SparkAppConfig$ - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.StatusUpdate - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.StatusUpdate$ - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.StopDriver$ - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.StopExecutor$ - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.StopExecutors$ - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.UpdateDelegationTokens - org.apache.spark.scheduler.cluster中的类
-
- CoarseGrainedClusterMessages.UpdateDelegationTokens$ - org.apache.spark.scheduler.cluster中的类
-
- code() - 类 中的方法org.apache.spark.mllib.feature.VocabWord
-
- CodegenMetrics - org.apache.spark.metrics.source中的类
-
:: Experimental ::
Metrics for code generation.
- CodegenMetrics() - 类 的构造器org.apache.spark.metrics.source.CodegenMetrics
-
- codeLen() - 类 中的方法org.apache.spark.mllib.feature.VocabWord
-
- coefficientMatrix() - 类 中的方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- coefficients() - 类 中的方法org.apache.spark.ml.classification.LinearSVCModel
-
- coefficients() - 类 中的方法org.apache.spark.ml.classification.LogisticRegressionModel
-
A vector of model coefficients for "binomial" logistic regression.
- coefficients() - 类 中的方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- coefficients() - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- coefficients() - 类 中的方法org.apache.spark.ml.regression.LinearRegressionModel
-
- coefficientStandardErrors() - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegressionTrainingSummary
-
Standard error of estimated coefficients and intercept.
- coefficientStandardErrors() - 类 中的方法org.apache.spark.ml.regression.LinearRegressionSummary
-
Standard error of estimated coefficients and intercept.
- cogroup(JavaPairRDD<K, W>, Partitioner) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
For each key k in this or other, return a resulting RDD that contains a tuple with the
list of values for that key in this as well as other.
- cogroup(JavaPairRDD<K, W1>, JavaPairRDD<K, W2>, Partitioner) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
For each key k in this or other1 or other2, return a resulting RDD that contains a
tuple with the list of values for that key in this, other1 and other2.
- cogroup(JavaPairRDD<K, W1>, JavaPairRDD<K, W2>, JavaPairRDD<K, W3>, Partitioner) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
For each key k in this or other1 or other2 or other3,
return a resulting RDD that contains a tuple with the list of values
for that key in this, other1, other2 and other3.
- cogroup(JavaPairRDD<K, W>) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
For each key k in this or other, return a resulting RDD that contains a tuple with the
list of values for that key in this as well as other.
- cogroup(JavaPairRDD<K, W1>, JavaPairRDD<K, W2>) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
For each key k in this or other1 or other2, return a resulting RDD that contains a
tuple with the list of values for that key in this, other1 and other2.
- cogroup(JavaPairRDD<K, W1>, JavaPairRDD<K, W2>, JavaPairRDD<K, W3>) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
For each key k in this or other1 or other2 or other3,
return a resulting RDD that contains a tuple with the list of values
for that key in this, other1, other2 and other3.
- cogroup(JavaPairRDD<K, W>, int) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
For each key k in this or other, return a resulting RDD that contains a tuple with the
list of values for that key in this as well as other.
- cogroup(JavaPairRDD<K, W1>, JavaPairRDD<K, W2>, int) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
For each key k in this or other1 or other2, return a resulting RDD that contains a
tuple with the list of values for that key in this, other1 and other2.
- cogroup(JavaPairRDD<K, W1>, JavaPairRDD<K, W2>, JavaPairRDD<K, W3>, int) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
For each key k in this or other1 or other2 or other3,
return a resulting RDD that contains a tuple with the list of values
for that key in this, other1, other2 and other3.
- cogroup(RDD<Tuple2<K, W1>>, RDD<Tuple2<K, W2>>, RDD<Tuple2<K, W3>>, Partitioner) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
For each key k in this or other1 or other2 or other3,
return a resulting RDD that contains a tuple with the list of values
for that key in this, other1, other2 and other3.
- cogroup(RDD<Tuple2<K, W>>, Partitioner) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
For each key k in this or other, return a resulting RDD that contains a tuple with the
list of values for that key in this as well as other.
- cogroup(RDD<Tuple2<K, W1>>, RDD<Tuple2<K, W2>>, Partitioner) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
For each key k in this or other1 or other2, return a resulting RDD that contains a
tuple with the list of values for that key in this, other1 and other2.
- cogroup(RDD<Tuple2<K, W1>>, RDD<Tuple2<K, W2>>, RDD<Tuple2<K, W3>>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
For each key k in this or other1 or other2 or other3,
return a resulting RDD that contains a tuple with the list of values
for that key in this, other1, other2 and other3.
- cogroup(RDD<Tuple2<K, W>>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
For each key k in this or other, return a resulting RDD that contains a tuple with the
list of values for that key in this as well as other.
- cogroup(RDD<Tuple2<K, W1>>, RDD<Tuple2<K, W2>>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
For each key k in this or other1 or other2, return a resulting RDD that contains a
tuple with the list of values for that key in this, other1 and other2.
- cogroup(RDD<Tuple2<K, W>>, int) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
For each key k in this or other, return a resulting RDD that contains a tuple with the
list of values for that key in this as well as other.
- cogroup(RDD<Tuple2<K, W1>>, RDD<Tuple2<K, W2>>, int) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
For each key k in this or other1 or other2, return a resulting RDD that contains a
tuple with the list of values for that key in this, other1 and other2.
- cogroup(RDD<Tuple2<K, W1>>, RDD<Tuple2<K, W2>>, RDD<Tuple2<K, W3>>, int) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
For each key k in this or other1 or other2 or other3,
return a resulting RDD that contains a tuple with the list of values
for that key in this, other1, other2 and other3.
- cogroup(KeyValueGroupedDataset<K, U>, Function3<K, Iterator<V>, Iterator<U>, TraversableOnce<R>>, Encoder<R>) - 类 中的方法org.apache.spark.sql.KeyValueGroupedDataset
-
(Scala-specific)
Applies the given function to each cogrouped data.
- cogroup(KeyValueGroupedDataset<K, U>, CoGroupFunction<K, V, U, R>, Encoder<R>) - 类 中的方法org.apache.spark.sql.KeyValueGroupedDataset
-
(Java-specific)
Applies the given function to each cogrouped data.
- cogroup(JavaPairDStream<K, W>) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying 'cogroup' between RDDs of this DStream and other DStream.
- cogroup(JavaPairDStream<K, W>, int) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying 'cogroup' between RDDs of this DStream and other DStream.
- cogroup(JavaPairDStream<K, W>, Partitioner) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying 'cogroup' between RDDs of this DStream and other DStream.
- cogroup(JavaPairDStream<K, W>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- cogroup(JavaPairDStream<K, W>, int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- cogroup(JavaPairDStream<K, W>, Partitioner) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- cogroup(JavaPairDStream<K, W>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- cogroup(JavaPairDStream<K, W>, int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- cogroup(JavaPairDStream<K, W>, Partitioner) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- cogroup(DStream<Tuple2<K, W>>, ClassTag<W>) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying 'cogroup' between RDDs of this DStream and other DStream.
- cogroup(DStream<Tuple2<K, W>>, int, ClassTag<W>) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying 'cogroup' between RDDs of this DStream and other DStream.
- cogroup(DStream<Tuple2<K, W>>, Partitioner, ClassTag<W>) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying 'cogroup' between RDDs of this DStream and other DStream.
- CoGroupedRDD<K> - org.apache.spark.rdd中的类
-
:: DeveloperApi ::
An RDD that cogroups its parents.
- CoGroupedRDD(Seq<RDD<? extends Product2<K, ?>>>, Partitioner, ClassTag<K>) - 类 的构造器org.apache.spark.rdd.CoGroupedRDD
-
- CoGroupFunction<K,V1,V2,R> - org.apache.spark.api.java.function中的接口
-
A function that returns zero or more output records from each grouping key and its values from 2
Datasets.
- col(String) - 类 中的方法org.apache.spark.sql.Dataset
-
Selects column based on the column name and returns it as a
Column.
- col(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns a
Column based on the given column name.
- coldStartStrategy() - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- coldStartStrategy() - 类 中的静态方法org.apache.spark.ml.recommendation.ALSModel
-
- colIter() - 类 中的方法org.apache.spark.ml.linalg.DenseMatrix
-
- colIter() - 接口 中的方法org.apache.spark.ml.linalg.Matrix
-
Returns an iterator of column vectors.
- colIter() - 类 中的方法org.apache.spark.ml.linalg.SparseMatrix
-
- colIter() - 类 中的方法org.apache.spark.mllib.linalg.DenseMatrix
-
- colIter() - 接口 中的方法org.apache.spark.mllib.linalg.Matrix
-
Returns an iterator of column vectors.
- colIter() - 类 中的方法org.apache.spark.mllib.linalg.SparseMatrix
-
- collect() - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- collect() - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- collect() - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- collect() - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Return an array that contains all of the elements in this RDD.
- collect() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- collect(PartialFunction<T, U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- collect() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- collect(PartialFunction<T, U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- collect() - 类 中的方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- collect() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- collect(PartialFunction<T, U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- collect() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- collect(PartialFunction<T, U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- collect() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- collect(PartialFunction<T, U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- collect() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- collect(PartialFunction<T, U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- collect() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- collect(PartialFunction<T, U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- collect() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- collect(PartialFunction<T, U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- collect() - 类 中的方法org.apache.spark.rdd.RDD
-
Return an array that contains all of the elements in this RDD.
- collect(PartialFunction<T, U>, ClassTag<U>) - 类 中的方法org.apache.spark.rdd.RDD
-
Return an RDD that contains all matching values by applying f.
- collect() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- collect(PartialFunction<T, U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- collect() - 类 中的方法org.apache.spark.sql.Dataset
-
Returns an array that contains all rows in this Dataset.
- collect(PartialFunction<BaseType, B>) - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- collect(PartialFunction<BaseType, B>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- collect(PartialFunction<BaseType, B>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- collect(PartialFunction<BaseType, B>) - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- collect(PartialFunction<A, B>, CanBuildFrom<Repr, B, That>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- collect_list(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns a list of objects with duplicates.
- collect_list(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns a list of objects with duplicates.
- collect_set(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns a set of objects with duplicate elements eliminated.
- collect_set(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns a set of objects with duplicate elements eliminated.
- collectAsList() - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a Java list that contains all rows in this Dataset.
- collectAsMap() - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Return the key-value pairs in this RDD to the master as a Map.
- collectAsMap() - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Return the key-value pairs in this RDD to the master as a Map.
- collectAsync() - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- collectAsync() - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- collectAsync() - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- collectAsync() - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
The asynchronous version of collect, which returns a future for
retrieving an array containing all of the elements in this RDD.
- collectAsync() - 类 中的方法org.apache.spark.rdd.AsyncRDDActions
-
Returns a future for retrieving all elements of this RDD.
- collectEdges(EdgeDirection) - 类 中的方法org.apache.spark.graphx.GraphOps
-
Returns an RDD that contains for each vertex v its local edges,
i.e., the edges that are incident on v, in the user-specified direction.
- collectFirst(PartialFunction<BaseType, B>) - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- collectFirst(PartialFunction<BaseType, B>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- collectFirst(PartialFunction<BaseType, B>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- collectFirst(PartialFunction<BaseType, B>) - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- collectFirst(PartialFunction<A, B>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- collectionAccumulator() - 类 中的方法org.apache.spark.SparkContext
-
Create and register a CollectionAccumulator, which starts with empty list and accumulates
inputs by adding them into the list.
- collectionAccumulator(String) - 类 中的方法org.apache.spark.SparkContext
-
Create and register a CollectionAccumulator, which starts with empty list and accumulates
inputs by adding them into the list.
- CollectionAccumulator<T> - org.apache.spark.util中的类
-
- CollectionAccumulator() - 类 的构造器org.apache.spark.util.CollectionAccumulator
-
- CollectionsUtils - org.apache.spark.util中的类
-
- CollectionsUtils() - 类 的构造器org.apache.spark.util.CollectionsUtils
-
- collectLeaves() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- collectLeaves() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- collectLeaves() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- collectLeaves() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- collectNeighborIds(EdgeDirection) - 类 中的方法org.apache.spark.graphx.GraphOps
-
Collect the neighbor vertex ids for each vertex.
- collectNeighbors(EdgeDirection) - 类 中的方法org.apache.spark.graphx.GraphOps
-
Collect the neighbor vertex attributes for each vertex.
- collectPartitions(int[]) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- collectPartitions(int[]) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- collectPartitions(int[]) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- collectPartitions(int[]) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Return an array that contains all of the elements in a specific partition of this RDD.
- collectSubModels() - 接口 中的方法org.apache.spark.ml.param.shared.HasCollectSubModels
-
Param for whether to collect a list of sub-models trained during tuning.
- collectSubModels() - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidator
-
- collectSubModels() - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplit
-
- colPtrs() - 类 中的方法org.apache.spark.ml.linalg.SparseMatrix
-
- colPtrs() - 类 中的方法org.apache.spark.mllib.linalg.SparseMatrix
-
- colRegex(String) - 类 中的方法org.apache.spark.sql.Dataset
-
Selects column based on the column name specified as a regex and returns it as
Column.
- colsPerBlock() - 类 中的方法org.apache.spark.mllib.linalg.distributed.BlockMatrix
-
- colStats(RDD<Vector>) - 类 中的静态方法org.apache.spark.mllib.stat.Statistics
-
Computes column-wise summary statistics for the input RDD[Vector].
- Column - org.apache.spark.sql.catalog中的类
-
A column in Spark, as returned by
listColumns method in
Catalog.
- Column(String, String, String, boolean, boolean, boolean) - 类 的构造器org.apache.spark.sql.catalog.Column
-
- Column - org.apache.spark.sql中的类
-
A column that will be computed based on the data in a DataFrame.
- Column(Expression) - 类 的构造器org.apache.spark.sql.Column
-
- Column(String) - 类 的构造器org.apache.spark.sql.Column
-
- column(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns a
Column based on the given column name.
- column(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarBatch
-
Returns the column at `ordinal`.
- ColumnarArray - org.apache.spark.sql.vectorized中的类
-
- ColumnarArray(ColumnVector, int, int) - 类 的构造器org.apache.spark.sql.vectorized.ColumnarArray
-
- ColumnarBatch - org.apache.spark.sql.vectorized中的类
-
This class wraps multiple ColumnVectors as a row-wise table.
- ColumnarBatch(ColumnVector[]) - 类 的构造器org.apache.spark.sql.vectorized.ColumnarBatch
-
- ColumnarMap - org.apache.spark.sql.vectorized中的类
-
- ColumnarMap(ColumnVector, ColumnVector, int, int) - 类 的构造器org.apache.spark.sql.vectorized.ColumnarMap
-
- ColumnarRow - org.apache.spark.sql.vectorized中的类
-
- ColumnarRow(ColumnVector, int) - 类 的构造器org.apache.spark.sql.vectorized.ColumnarRow
-
- ColumnName - org.apache.spark.sql中的类
-
A convenient class used for constructing schema.
- ColumnName(String) - 类 的构造器org.apache.spark.sql.ColumnName
-
- ColumnPruner - org.apache.spark.ml.feature中的类
-
Utility transformer for removing temporary columns from a DataFrame.
- ColumnPruner(String, Set<String>) - 类 的构造器org.apache.spark.ml.feature.ColumnPruner
-
- ColumnPruner(Set<String>) - 类 的构造器org.apache.spark.ml.feature.ColumnPruner
-
- columns() - 类 中的方法org.apache.spark.sql.Dataset
-
Returns all column names as an array.
- columnSchema() - 类 中的静态方法org.apache.spark.ml.image.ImageSchema
-
Schema for the image column: Row(String, Int, Int, Int, Int, Array[Byte])
- columnSimilarities() - 类 中的方法org.apache.spark.mllib.linalg.distributed.IndexedRowMatrix
-
Compute all cosine similarities between columns of this matrix using the brute-force
approach of computing normalized dot products.
- columnSimilarities() - 类 中的方法org.apache.spark.mllib.linalg.distributed.RowMatrix
-
Compute all cosine similarities between columns of this matrix using the brute-force
approach of computing normalized dot products.
- columnSimilarities(double) - 类 中的方法org.apache.spark.mllib.linalg.distributed.RowMatrix
-
Compute similarities between columns of this matrix using a sampling approach.
- columnsToPrune() - 类 中的方法org.apache.spark.ml.feature.ColumnPruner
-
- ColumnVector - org.apache.spark.sql.vectorized中的类
-
An interface representing in-memory columnar data in Spark.
- combinations(int) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- combineByKey(Function<V, C>, Function2<C, V, C>, Function2<C, C, C>, Partitioner, boolean, Serializer) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Generic function to combine the elements for each key using a custom set of aggregation
functions.
- combineByKey(Function<V, C>, Function2<C, V, C>, Function2<C, C, C>, Partitioner) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Generic function to combine the elements for each key using a custom set of aggregation
functions.
- combineByKey(Function<V, C>, Function2<C, V, C>, Function2<C, C, C>, int) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Simplified version of combineByKey that hash-partitions the output RDD and uses map-side
aggregation.
- combineByKey(Function<V, C>, Function2<C, V, C>, Function2<C, C, C>) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Simplified version of combineByKey that hash-partitions the resulting RDD using the existing
partitioner/parallelism level and using map-side aggregation.
- combineByKey(Function1<V, C>, Function2<C, V, C>, Function2<C, C, C>, Partitioner, boolean, Serializer) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Generic function to combine the elements for each key using a custom set of aggregation
functions.
- combineByKey(Function1<V, C>, Function2<C, V, C>, Function2<C, C, C>, int) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Simplified version of combineByKeyWithClassTag that hash-partitions the output RDD.
- combineByKey(Function1<V, C>, Function2<C, V, C>, Function2<C, C, C>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Simplified version of combineByKeyWithClassTag that hash-partitions the resulting RDD using the
existing partitioner/parallelism level.
- combineByKey(Function<V, C>, Function2<C, V, C>, Function2<C, C, C>, Partitioner) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Combine elements of each key in DStream's RDDs using custom function.
- combineByKey(Function<V, C>, Function2<C, V, C>, Function2<C, C, C>, Partitioner, boolean) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Combine elements of each key in DStream's RDDs using custom function.
- combineByKey(Function<V, C>, Function2<C, V, C>, Function2<C, C, C>, Partitioner) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- combineByKey(Function<V, C>, Function2<C, V, C>, Function2<C, C, C>, Partitioner, boolean) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- combineByKey(Function<V, C>, Function2<C, V, C>, Function2<C, C, C>, Partitioner) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- combineByKey(Function<V, C>, Function2<C, V, C>, Function2<C, C, C>, Partitioner, boolean) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- combineByKey(Function1<V, C>, Function2<C, V, C>, Function2<C, C, C>, Partitioner, boolean, ClassTag<C>) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Combine elements of each key in DStream's RDDs using custom functions.
- combineByKeyWithClassTag(Function1<V, C>, Function2<C, V, C>, Function2<C, C, C>, Partitioner, boolean, Serializer, ClassTag<C>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
:: Experimental ::
Generic function to combine the elements for each key using a custom set of aggregation
functions.
- combineByKeyWithClassTag(Function1<V, C>, Function2<C, V, C>, Function2<C, C, C>, int, ClassTag<C>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
:: Experimental ::
Simplified version of combineByKeyWithClassTag that hash-partitions the output RDD.
- combineByKeyWithClassTag(Function1<V, C>, Function2<C, V, C>, Function2<C, C, C>, ClassTag<C>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
:: Experimental ::
Simplified version of combineByKeyWithClassTag that hash-partitions the resulting RDD using the
existing partitioner/parallelism level.
- combineCombinersByKey(Iterator<? extends Product2<K, C>>, TaskContext) - 类 中的方法org.apache.spark.Aggregator
-
- combineValuesByKey(Iterator<? extends Product2<K, V>>, TaskContext) - 类 中的方法org.apache.spark.Aggregator
-
- commit(Function0<Parsers.Parser<T>>) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaParser
-
- commit(Offset) - 接口 中的方法org.apache.spark.sql.sources.v2.reader.streaming.ContinuousReader
-
Informs the source that Spark has completed processing all data for offsets less than or
equal to `end` and will only request offsets greater than `end` in the future.
- commit(Offset) - 接口 中的方法org.apache.spark.sql.sources.v2.reader.streaming.MicroBatchReader
-
Informs the source that Spark has completed processing all data for offsets less than or
equal to `end` and will only request offsets greater than `end` in the future.
- commit(WriterCommitMessage[]) - 接口 中的方法org.apache.spark.sql.sources.v2.writer.DataSourceWriter
-
Commits this writing job with a list of commit messages.
- commit() - 接口 中的方法org.apache.spark.sql.sources.v2.writer.DataWriter
-
- commit(long, WriterCommitMessage[]) - 接口 中的方法org.apache.spark.sql.sources.v2.writer.streaming.StreamWriter
-
Commits this writing job for the specified epoch with a list of commit messages.
- commit(WriterCommitMessage[]) - 接口 中的方法org.apache.spark.sql.sources.v2.writer.streaming.StreamWriter
-
- commitJob(JobContext, Seq<FileCommitProtocol.TaskCommitMessage>) - 类 中的方法org.apache.spark.internal.io.FileCommitProtocol
-
Commits a job after the writes succeed.
- commitJob(JobContext, Seq<FileCommitProtocol.TaskCommitMessage>) - 类 中的方法org.apache.spark.internal.io.HadoopMapReduceCommitProtocol
-
- commitTask(TaskAttemptContext) - 类 中的方法org.apache.spark.internal.io.FileCommitProtocol
-
Commits a task after the writes succeed.
- commitTask(TaskAttemptContext) - 类 中的方法org.apache.spark.internal.io.HadoopMapReduceCommitProtocol
-
- commitTask(OutputCommitter, TaskAttemptContext, int, int) - 类 中的静态方法org.apache.spark.mapred.SparkHadoopMapRedUtil
-
Commits a task output.
- commonHeaderNodes() - 类 中的静态方法org.apache.spark.ui.UIUtils
-
- companion() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- compare(PartitionGroup, PartitionGroup) - 类 中的方法org.apache.spark.rdd.DefaultPartitionCoalescer
-
- compare(Option<PartitionGroup>, Option<PartitionGroup>) - 类 中的方法org.apache.spark.rdd.DefaultPartitionCoalescer
-
- compare(Decimal) - 类 中的方法org.apache.spark.sql.types.Decimal
-
- compare(RDDInfo) - 类 中的方法org.apache.spark.storage.RDDInfo
-
- compareTo(A) - 类 中的静态方法org.apache.spark.sql.types.Decimal
-
- compareTo(A) - 类 中的静态方法org.apache.spark.storage.RDDInfo
-
- compareTo(SparkShutdownHook) - 类 中的方法org.apache.spark.util.SparkShutdownHook
-
- compileValue(Object) - 类 中的静态方法org.apache.spark.sql.jdbc.DB2Dialect
-
- compileValue(Object) - 类 中的静态方法org.apache.spark.sql.jdbc.DerbyDialect
-
- compileValue(Object) - 类 中的方法org.apache.spark.sql.jdbc.JdbcDialect
-
Converts value to SQL expression.
- compileValue(Object) - 类 中的静态方法org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- compileValue(Object) - 类 中的静态方法org.apache.spark.sql.jdbc.MySQLDialect
-
- compileValue(Object) - 类 中的静态方法org.apache.spark.sql.jdbc.NoopDialect
-
- compileValue(Object) - 类 中的静态方法org.apache.spark.sql.jdbc.OracleDialect
-
- compileValue(Object) - 类 中的静态方法org.apache.spark.sql.jdbc.PostgresDialect
-
- compileValue(Object) - 类 中的静态方法org.apache.spark.sql.jdbc.TeradataDialect
-
- Complete() - 类 中的静态方法org.apache.spark.sql.streaming.OutputMode
-
OutputMode in which all the rows in the streaming DataFrame/Dataset will be written
to the sink every time there are some updates.
- completed() - 类 中的方法org.apache.spark.status.api.v1.ApplicationAttemptInfo
-
- completedIndices() - 类 中的方法org.apache.spark.status.LiveJob
-
- completedIndices() - 类 中的方法org.apache.spark.status.LiveStage
-
- completedStages() - 类 中的方法org.apache.spark.status.LiveJob
-
- completedTasks() - 类 中的方法org.apache.spark.status.api.v1.ExecutorSummary
-
- completedTasks() - 类 中的方法org.apache.spark.status.LiveExecutor
-
- completedTasks() - 类 中的方法org.apache.spark.status.LiveJob
-
- completedTasks() - 类 中的方法org.apache.spark.status.LiveStage
-
- COMPLETION_TIME() - 类 中的静态方法org.apache.spark.status.TaskIndexNames
-
- completionTime() - 类 中的方法org.apache.spark.scheduler.StageInfo
-
Time when all tasks in the stage completed or when the stage was cancelled.
- completionTime() - 类 中的方法org.apache.spark.status.api.v1.JobData
-
- completionTime() - 类 中的方法org.apache.spark.status.api.v1.StageData
-
- completionTime() - 类 中的方法org.apache.spark.status.LiveJob
-
- ComplexFutureAction<T> - org.apache.spark中的类
-
A
FutureAction for actions that could trigger multiple Spark jobs.
- ComplexFutureAction(Function1<JobSubmitter, Future<T>>) - 类 的构造器org.apache.spark.ComplexFutureAction
-
- compose(Function1<A, T1>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- compressed() - 类 中的静态方法org.apache.spark.ml.linalg.DenseMatrix
-
- compressed() - 类 中的静态方法org.apache.spark.ml.linalg.DenseVector
-
- compressed() - 接口 中的方法org.apache.spark.ml.linalg.Matrix
-
Returns a matrix in dense column major, dense row major, sparse row major, or sparse column
major format, whichever uses less storage.
- compressed() - 类 中的静态方法org.apache.spark.ml.linalg.SparseMatrix
-
- compressed() - 类 中的静态方法org.apache.spark.ml.linalg.SparseVector
-
- compressed() - 接口 中的方法org.apache.spark.ml.linalg.Vector
-
Returns a vector in either dense or sparse format, whichever uses less storage.
- compressed() - 类 中的静态方法org.apache.spark.mllib.linalg.DenseVector
-
- compressed() - 类 中的静态方法org.apache.spark.mllib.linalg.SparseVector
-
- compressed() - 接口 中的方法org.apache.spark.mllib.linalg.Vector
-
Returns a vector in either dense or sparse format, whichever uses less storage.
- compressedColMajor() - 类 中的静态方法org.apache.spark.ml.linalg.DenseMatrix
-
- compressedColMajor() - 接口 中的方法org.apache.spark.ml.linalg.Matrix
-
Returns a matrix in dense or sparse column major format, whichever uses less storage.
- compressedColMajor() - 类 中的静态方法org.apache.spark.ml.linalg.SparseMatrix
-
- compressedInputStream(InputStream) - 接口 中的方法org.apache.spark.io.CompressionCodec
-
- compressedInputStream(InputStream) - 类 中的方法org.apache.spark.io.LZ4CompressionCodec
-
- compressedInputStream(InputStream) - 类 中的方法org.apache.spark.io.LZFCompressionCodec
-
- compressedInputStream(InputStream) - 类 中的方法org.apache.spark.io.SnappyCompressionCodec
-
- compressedInputStream(InputStream) - 类 中的方法org.apache.spark.io.ZStdCompressionCodec
-
- compressedOutputStream(OutputStream) - 接口 中的方法org.apache.spark.io.CompressionCodec
-
- compressedOutputStream(OutputStream) - 类 中的方法org.apache.spark.io.LZ4CompressionCodec
-
- compressedOutputStream(OutputStream) - 类 中的方法org.apache.spark.io.LZFCompressionCodec
-
- compressedOutputStream(OutputStream) - 类 中的方法org.apache.spark.io.SnappyCompressionCodec
-
- compressedOutputStream(OutputStream) - 类 中的方法org.apache.spark.io.ZStdCompressionCodec
-
- compressedRowMajor() - 类 中的静态方法org.apache.spark.ml.linalg.DenseMatrix
-
- compressedRowMajor() - 接口 中的方法org.apache.spark.ml.linalg.Matrix
-
Returns a matrix in dense or sparse row major format, whichever uses less storage.
- compressedRowMajor() - 类 中的静态方法org.apache.spark.ml.linalg.SparseMatrix
-
- CompressionCodec - org.apache.spark.io中的接口
-
:: DeveloperApi ::
CompressionCodec allows the customization of choosing different compression implementations
to be used in block storage.
- compute(Partition, TaskContext) - 类 中的方法org.apache.spark.api.r.BaseRRDD
-
- compute(Partition, TaskContext) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- compute(Partition, TaskContext) - 类 中的方法org.apache.spark.graphx.EdgeRDD
-
- compute(Partition, TaskContext) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- compute(Partition, TaskContext) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- compute(Partition, TaskContext) - 类 中的方法org.apache.spark.graphx.VertexRDD
-
Provides the RDD[(VertexId, VD)] equivalent output.
- compute(Vector, double, Vector) - 类 中的方法org.apache.spark.mllib.optimization.Gradient
-
Compute the gradient and loss given the features of a single data point.
- compute(Vector, double, Vector, Vector) - 类 中的方法org.apache.spark.mllib.optimization.Gradient
-
Compute the gradient and loss given the features of a single data point,
add the gradient to a provided vector to avoid creating new objects, and return loss.
- compute(Vector, double, Vector) - 类 中的方法org.apache.spark.mllib.optimization.HingeGradient
-
- compute(Vector, double, Vector, Vector) - 类 中的方法org.apache.spark.mllib.optimization.HingeGradient
-
- compute(Vector, Vector, double, int, double) - 类 中的方法org.apache.spark.mllib.optimization.L1Updater
-
- compute(Vector, double, Vector) - 类 中的方法org.apache.spark.mllib.optimization.LeastSquaresGradient
-
- compute(Vector, double, Vector, Vector) - 类 中的方法org.apache.spark.mllib.optimization.LeastSquaresGradient
-
- compute(Vector, double, Vector, Vector) - 类 中的方法org.apache.spark.mllib.optimization.LogisticGradient
-
- compute(Vector, Vector, double, int, double) - 类 中的方法org.apache.spark.mllib.optimization.SimpleUpdater
-
- compute(Vector, Vector, double, int, double) - 类 中的方法org.apache.spark.mllib.optimization.SquaredL2Updater
-
- compute(Vector, Vector, double, int, double) - 类 中的方法org.apache.spark.mllib.optimization.Updater
-
Compute an updated value for weights given the gradient, stepSize, iteration number and
regularization parameter.
- compute(Partition, TaskContext) - 类 中的方法org.apache.spark.rdd.CoGroupedRDD
-
- compute(Partition, TaskContext) - 类 中的方法org.apache.spark.rdd.HadoopRDD
-
- compute(Partition, TaskContext) - 类 中的方法org.apache.spark.rdd.JdbcRDD
-
- compute(Partition, TaskContext) - 类 中的方法org.apache.spark.rdd.NewHadoopRDD
-
- compute(Partition, TaskContext) - 类 中的方法org.apache.spark.rdd.PartitionPruningRDD
-
- compute(Partition, TaskContext) - 类 中的方法org.apache.spark.rdd.RDD
-
:: DeveloperApi ::
Implemented by subclasses to compute a given partition.
- compute(Partition, TaskContext) - 类 中的方法org.apache.spark.rdd.ShuffledRDD
-
- compute(Partition, TaskContext) - 类 中的方法org.apache.spark.rdd.UnionRDD
-
- compute(Time) - 类 中的方法org.apache.spark.streaming.api.java.JavaDStream
-
Generate an RDD for the given duration
- compute(Time) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- compute(Time) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Method that generates an RDD for the given Duration
- compute(Time) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- compute(Time) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- compute(Time) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- compute(Time) - 类 中的方法org.apache.spark.streaming.dstream.ConstantInputDStream
-
- compute(Time) - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Method that generates an RDD for the given time
- compute(Time) - 类 中的方法org.apache.spark.streaming.dstream.ReceiverInputDStream
-
- computeClusterStats(Dataset<Row>, String, String) - 类 中的静态方法org.apache.spark.ml.evaluation.SquaredEuclideanSilhouette
-
The method takes the input dataset and computes the aggregated values
about a cluster which are needed by the algorithm.
- computeColumnSummaryStatistics() - 类 中的方法org.apache.spark.mllib.linalg.distributed.RowMatrix
-
Computes column-wise summary statistics.
- computeCorrelation(RDD<Object>, RDD<Object>) - 类 中的静态方法org.apache.spark.mllib.stat.correlation.PearsonCorrelation
-
Compute the Pearson correlation for two datasets.
- computeCorrelation(RDD<Object>, RDD<Object>) - 类 中的静态方法org.apache.spark.mllib.stat.correlation.SpearmanCorrelation
-
Compute Spearman's correlation for two datasets.
- computeCorrelationMatrix(RDD<Vector>) - 类 中的静态方法org.apache.spark.mllib.stat.correlation.PearsonCorrelation
-
Compute the Pearson correlation matrix S, for the input matrix, where S(i, j) is the
correlation between column i and j. 0 covariance results in a correlation value of Double.NaN.
- computeCorrelationMatrix(RDD<Vector>) - 类 中的静态方法org.apache.spark.mllib.stat.correlation.SpearmanCorrelation
-
Compute Spearman's correlation matrix S, for the input matrix, where S(i, j) is the
correlation between column i and j.
- computeCorrelationMatrixFromCovariance(Matrix) - 类 中的静态方法org.apache.spark.mllib.stat.correlation.PearsonCorrelation
-
Compute the Pearson correlation matrix from the covariance matrix.
0 variance results in a correlation value of Double.NaN.
- computeCorrelationWithMatrixImpl(RDD<Object>, RDD<Object>) - 类 中的静态方法org.apache.spark.mllib.stat.correlation.PearsonCorrelation
-
- computeCorrelationWithMatrixImpl(RDD<Object>, RDD<Object>) - 类 中的静态方法org.apache.spark.mllib.stat.correlation.SpearmanCorrelation
-
- computeCost(Dataset<?>) - 类 中的方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
Computes the sum of squared distances between the input points and their corresponding cluster
centers.
- computeCost(Dataset<?>) - 类 中的方法org.apache.spark.ml.clustering.KMeansModel
-
Return the K-means cost (sum of squared distances of points to their nearest center) for this
model on the given data.
- computeCost(Vector) - 类 中的方法org.apache.spark.mllib.clustering.BisectingKMeansModel
-
Computes the squared distance between the input point and the cluster center it belongs to.
- computeCost(RDD<Vector>) - 类 中的方法org.apache.spark.mllib.clustering.BisectingKMeansModel
-
Computes the sum of squared distances between the input points and their corresponding cluster
centers.
- computeCost(JavaRDD<Vector>) - 类 中的方法org.apache.spark.mllib.clustering.BisectingKMeansModel
-
Java-friendly version of computeCost().
- computeCost(RDD<Vector>) - 类 中的方法org.apache.spark.mllib.clustering.KMeansModel
-
Return the K-means cost (sum of squared distances of points to their nearest center) for this
model on the given data.
- computeCovariance() - 类 中的方法org.apache.spark.mllib.linalg.distributed.RowMatrix
-
Computes the covariance matrix, treating each row as an observation.
- computeError(RDD<LabeledPoint>, DecisionTreeRegressionModel[], double[], Loss) - 类 中的静态方法org.apache.spark.ml.tree.impl.GradientBoostedTrees
-
Method to calculate error of the base learner for the gradient boosting calculation.
- computeError(org.apache.spark.mllib.tree.model.TreeEnsembleModel, RDD<LabeledPoint>) - 接口 中的方法org.apache.spark.mllib.tree.loss.Loss
-
Method to calculate error of the base learner for the gradient boosting calculation.
- computeError(double, double) - 接口 中的方法org.apache.spark.mllib.tree.loss.Loss
-
Method to calculate loss when the predictions are already known.
- computeFractionForSampleSize(int, long, boolean) - 类 中的静态方法org.apache.spark.util.random.SamplingUtils
-
Returns a sampling rate that guarantees a sample of size greater than or equal to
sampleSizeLowerBound 99.99% of the time.
- computeGramianMatrix() - 类 中的方法org.apache.spark.mllib.linalg.distributed.IndexedRowMatrix
-
Computes the Gramian matrix A^T A.
- computeGramianMatrix() - 类 中的方法org.apache.spark.mllib.linalg.distributed.RowMatrix
-
Computes the Gramian matrix A^T A.
- computeInitialPredictionAndError(RDD<LabeledPoint>, double, DecisionTreeRegressionModel, Loss) - 类 中的静态方法org.apache.spark.ml.tree.impl.GradientBoostedTrees
-
Compute the initial predictions and errors for a dataset for the first
iteration of gradient boosting.
- computeInitialPredictionAndError(RDD<LabeledPoint>, double, DecisionTreeModel, Loss) - 类 中的静态方法org.apache.spark.mllib.tree.model.GradientBoostedTreesModel
-
:: DeveloperApi ::
Compute the initial predictions and errors for a dataset for the first
iteration of gradient boosting.
- computePreferredLocations(Seq<InputFormatInfo>) - 类 中的静态方法org.apache.spark.scheduler.InputFormatInfo
-
Computes the preferred locations based on input(s) and returned a location to block map.
- computePrincipalComponents(int) - 类 中的方法org.apache.spark.mllib.linalg.distributed.RowMatrix
-
Computes the top k principal components only.
- computePrincipalComponentsAndExplainedVariance(int) - 类 中的方法org.apache.spark.mllib.linalg.distributed.RowMatrix
-
Computes the top k principal components and a vector of proportions of
variance explained by each principal component.
- computeSilhouetteCoefficient(Broadcast<Map<Object, SquaredEuclideanSilhouette.ClusterStats>>, Vector, double, double) - 类 中的静态方法org.apache.spark.ml.evaluation.SquaredEuclideanSilhouette
-
It computes the Silhouette coefficient for a point.
- computeSilhouetteScore(Dataset<?>, String, String) - 类 中的静态方法org.apache.spark.ml.evaluation.SquaredEuclideanSilhouette
-
Compute the mean Silhouette values of all samples.
- computeSVD(int, boolean, double) - 类 中的方法org.apache.spark.mllib.linalg.distributed.IndexedRowMatrix
-
Computes the singular value decomposition of this IndexedRowMatrix.
- computeSVD(int, boolean, double) - 类 中的方法org.apache.spark.mllib.linalg.distributed.RowMatrix
-
Computes singular value decomposition of this matrix.
- computeThresholdByKey(Map<K, AcceptanceResult>, Map<K, Object>) - 类 中的静态方法org.apache.spark.util.random.StratifiedSamplingUtils
-
Given the result returned by getCounts, determine the threshold for accepting items to
generate exact sample size.
- concat(Column...) - 类 中的静态方法org.apache.spark.sql.functions
-
Concatenates multiple input columns together into a single column.
- concat(Seq<Column>) - 类 中的静态方法org.apache.spark.sql.functions
-
Concatenates multiple input columns together into a single column.
- concat_ws(String, Column...) - 类 中的静态方法org.apache.spark.sql.functions
-
Concatenates multiple input string columns together into a single string column,
using the given separator.
- concat_ws(String, Seq<Column>) - 类 中的静态方法org.apache.spark.sql.functions
-
Concatenates multiple input string columns together into a single string column,
using the given separator.
- Conf(int, int, double, double, double, double, double, double) - 类 的构造器org.apache.spark.graphx.lib.SVDPlusPlus.Conf
-
- conf() - 类 中的方法org.apache.spark.SparkEnv
-
- conf() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- conf() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- conf() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- conf() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- conf() - 类 中的方法org.apache.spark.sql.hive.RelationConversions
-
- conf() - 类 中的方法org.apache.spark.sql.SparkSession
-
Runtime configuration interface for Spark.
- confidence() - 类 中的方法org.apache.spark.mllib.fpm.AssociationRules.Rule
-
Returns the confidence of the rule.
- confidence() - 类 中的方法org.apache.spark.partial.BoundedDouble
-
- confidence() - 类 中的方法org.apache.spark.util.sketch.CountMinSketch
-
- config(String, String) - 类 中的方法org.apache.spark.sql.SparkSession.Builder
-
Sets a config option.
- config(String, long) - 类 中的方法org.apache.spark.sql.SparkSession.Builder
-
Sets a config option.
- config(String, double) - 类 中的方法org.apache.spark.sql.SparkSession.Builder
-
Sets a config option.
- config(String, boolean) - 类 中的方法org.apache.spark.sql.SparkSession.Builder
-
Sets a config option.
- config(SparkConf) - 类 中的方法org.apache.spark.sql.SparkSession.Builder
-
Sets a list of config options based on the given SparkConf.
- config - org.apache.spark.status中的类
-
- config() - 类 的构造器org.apache.spark.status.config
-
- ConfigEntryWithDefault<T> - org.apache.spark.internal.config中的类
-
- ConfigEntryWithDefault(String, List<String>, T, Function1<String, T>, Function1<T, String>, String, boolean) - 类 的构造器org.apache.spark.internal.config.ConfigEntryWithDefault
-
- ConfigEntryWithDefaultFunction<T> - org.apache.spark.internal.config中的类
-
- ConfigEntryWithDefaultFunction(String, List<String>, Function0<T>, Function1<String, T>, Function1<T, String>, String, boolean) - 类 的构造器org.apache.spark.internal.config.ConfigEntryWithDefaultFunction
-
- ConfigEntryWithDefaultString<T> - org.apache.spark.internal.config中的类
-
- ConfigEntryWithDefaultString(String, List<String>, String, Function1<String, T>, Function1<T, String>, String, boolean) - 类 的构造器org.apache.spark.internal.config.ConfigEntryWithDefaultString
-
- ConfigHelpers - org.apache.spark.internal.config中的类
-
- ConfigHelpers() - 类 的构造器org.apache.spark.internal.config.ConfigHelpers
-
- configTestLog4j(String) - 类 中的静态方法org.apache.spark.util.Utils
-
config a log4j properties used for testsuite
- configuration() - 类 中的方法org.apache.spark.scheduler.InputFormatInfo
-
- CONFIGURATION_INSTANTIATION_LOCK() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
Configuration's constructor is not threadsafe (see SPARK-1097 and HADOOP-10456).
- CONFIGURATION_INSTANTIATION_LOCK() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
Configuration's constructor is not threadsafe (see SPARK-1097 and HADOOP-10456).
- configureJobPropertiesForStorageHandler(TableDesc, Configuration, boolean) - 类 中的静态方法org.apache.spark.sql.hive.HiveTableUtil
-
- confusionMatrix() - 类 中的方法org.apache.spark.mllib.evaluation.MulticlassMetrics
-
Returns confusion matrix:
predicted classes are in columns,
they are ordered by class label ascending,
as in "labels"
- connectedComponents() - 类 中的方法org.apache.spark.graphx.GraphOps
-
Compute the connected component membership of each vertex and return a graph with the vertex
value containing the lowest vertex id in the connected component containing that vertex.
- connectedComponents(int) - 类 中的方法org.apache.spark.graphx.GraphOps
-
Compute the connected component membership of each vertex and return a graph with the vertex
value containing the lowest vertex id in the connected component containing that vertex.
- ConnectedComponents - org.apache.spark.graphx.lib中的类
-
Connected components algorithm.
- ConnectedComponents() - 类 的构造器org.apache.spark.graphx.lib.ConnectedComponents
-
- consequent() - 类 中的方法org.apache.spark.mllib.fpm.AssociationRules.Rule
-
- ConstantInputDStream<T> - org.apache.spark.streaming.dstream中的类
-
An input stream that always returns the same RDD on each time step.
- ConstantInputDStream(StreamingContext, RDD<T>, ClassTag<T>) - 类 的构造器org.apache.spark.streaming.dstream.ConstantInputDStream
-
- constraints() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- constraints() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- constraints() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- constructTree(org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0.NodeData[]) - 类 中的方法org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$
-
Given a list of nodes from a tree, construct the tree.
- constructTrees(RDD<org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0.NodeData>) - 类 中的方法org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$
-
- constructURIForAuthentication(URI, org.apache.spark.SecurityManager) - 类 中的静态方法org.apache.spark.util.Utils
-
Construct a URI container information used for authentication.
- contains(Param<?>) - 类 中的方法org.apache.spark.ml.param.ParamMap
-
Checks whether a parameter is explicitly specified.
- contains(String) - 类 中的方法org.apache.spark.SparkConf
-
Does the configuration contain a given parameter?
- contains(Object) - 类 中的方法org.apache.spark.sql.Column
-
Contains the other element.
- contains(String) - 类 中的方法org.apache.spark.sql.types.Metadata
-
Tests whether this Metadata contains a binding for a key.
- contains(A1) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- containsBlock(BlockId) - 类 中的方法org.apache.spark.storage.StorageStatus
-
已过时。
Return whether the given block is stored in this block manager in O(1) time.
- containsChild() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- containsChild() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- containsChild() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- containsChild() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- containsDelimiters() - 类 中的方法org.apache.spark.sql.hive.execution.HiveOptions
-
- containsKey(Object) - 类 中的方法org.apache.spark.api.java.JavaUtils.SerializableMapWrapper
-
- containsNull() - 类 中的方法org.apache.spark.sql.types.ArrayType
-
- containsSlice(GenSeq<B>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- contentType() - 类 中的方法org.apache.spark.ui.JettyUtils.ServletParams
-
- context() - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- context() - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- context() - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- context() - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
- context() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- context() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- context() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- context() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- context() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- context() - 类 中的方法org.apache.spark.InterruptibleIterator
-
- context(SQLContext) - 类 中的静态方法org.apache.spark.ml.r.RWrappers
-
- context(SQLContext) - 类 中的方法org.apache.spark.ml.util.MLReader
-
- context(SQLContext) - 类 中的方法org.apache.spark.ml.util.MLWriter
-
- context() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- context() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- context() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- context() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- context() - 类 中的方法org.apache.spark.rdd.RDD
-
- context() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- context() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- context() - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
- context() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- context() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- context() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- context() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- context() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- context() - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Return the StreamingContext associated with this DStream
- Continuous() - 类 中的静态方法org.apache.spark.mllib.tree.configuration.FeatureType
-
- Continuous(long) - 类 中的静态方法org.apache.spark.sql.streaming.Trigger
-
A trigger that continuously processes streaming data, asynchronously checkpointing at
the specified interval.
- Continuous(long, TimeUnit) - 类 中的静态方法org.apache.spark.sql.streaming.Trigger
-
A trigger that continuously processes streaming data, asynchronously checkpointing at
the specified interval.
{{{
import java.util.concurrent.TimeUnit
df.writeStream.trigger(Trigger.Continuous(10, TimeUnit.SECONDS))
}}}
- Continuous(Duration) - 类 中的静态方法org.apache.spark.sql.streaming.Trigger
-
(Scala-friendly)
A trigger that continuously processes streaming data, asynchronously checkpointing at
the specified interval.
{{{
import scala.concurrent.duration._
df.writeStream.trigger(Trigger.Continuous(10.seconds))
}}}
- Continuous(String) - 类 中的静态方法org.apache.spark.sql.streaming.Trigger
-
A trigger that continuously processes streaming data, asynchronously checkpointing at
the specified interval.
{{{
df.writeStream.trigger(Trigger.Continuous("10 seconds"))
}}}
- ContinuousDataReader<T> - org.apache.spark.sql.sources.v2.reader.streaming中的接口
-
A variation on
DataReader for use with streaming in continuous processing mode.
- ContinuousReader - org.apache.spark.sql.sources.v2.reader.streaming中的接口
-
- ContinuousReadSupport - org.apache.spark.sql.sources.v2中的接口
-
- ContinuousSplit - org.apache.spark.ml.tree中的类
-
Split which tests a continuous feature.
- conv(Column, int, int) - 类 中的静态方法org.apache.spark.sql.functions
-
Convert a number in a string column from one base to another.
- CONVERT_METASTORE_ORC() - 类 中的静态方法org.apache.spark.sql.hive.HiveUtils
-
- CONVERT_METASTORE_PARQUET() - 类 中的静态方法org.apache.spark.sql.hive.HiveUtils
-
- CONVERT_METASTORE_PARQUET_WITH_SCHEMA_MERGING() - 类 中的静态方法org.apache.spark.sql.hive.HiveUtils
-
- convertMatrixColumnsFromML(Dataset<?>, String...) - 类 中的静态方法org.apache.spark.mllib.util.MLUtils
-
Converts matrix columns in an input Dataset to the
Matrix
type from the new
Matrix type under the
spark.ml package.
- convertMatrixColumnsFromML(Dataset<?>, Seq<String>) - 类 中的静态方法org.apache.spark.mllib.util.MLUtils
-
Converts matrix columns in an input Dataset to the
Matrix
type from the new
Matrix type under the
spark.ml package.
- convertMatrixColumnsToML(Dataset<?>, String...) - 类 中的静态方法org.apache.spark.mllib.util.MLUtils
-
Converts Matrix columns in an input Dataset from the
Matrix
type to the new
Matrix type under the
spark.ml package.
- convertMatrixColumnsToML(Dataset<?>, Seq<String>) - 类 中的静态方法org.apache.spark.mllib.util.MLUtils
-
Converts Matrix columns in an input Dataset from the
Matrix
type to the new
Matrix type under the
spark.ml package.
- convertToCanonicalEdges(Function2<ED, ED, ED>) - 类 中的方法org.apache.spark.graphx.GraphOps
-
Convert bi-directional edges into uni-directional ones.
- convertToTimeUnit(long, TimeUnit) - 类 中的静态方法org.apache.spark.streaming.ui.UIUtils
-
Convert milliseconds to the specified unit.
- convertVectorColumnsFromML(Dataset<?>, String...) - 类 中的静态方法org.apache.spark.mllib.util.MLUtils
-
Converts vector columns in an input Dataset to the
Vector
type from the new
Vector type under the
spark.ml package.
- convertVectorColumnsFromML(Dataset<?>, Seq<String>) - 类 中的静态方法org.apache.spark.mllib.util.MLUtils
-
Converts vector columns in an input Dataset to the
Vector
type from the new
Vector type under the
spark.ml package.
- convertVectorColumnsToML(Dataset<?>, String...) - 类 中的静态方法org.apache.spark.mllib.util.MLUtils
-
Converts vector columns in an input Dataset from the
Vector
type to the new
Vector type under the
spark.ml package.
- convertVectorColumnsToML(Dataset<?>, Seq<String>) - 类 中的静态方法org.apache.spark.mllib.util.MLUtils
-
Converts vector columns in an input Dataset from the
Vector
type to the new
Vector type under the
spark.ml package.
- CoordinateMatrix - org.apache.spark.mllib.linalg.distributed中的类
-
Represents a matrix in coordinate format.
- CoordinateMatrix(RDD<MatrixEntry>, long, long) - 类 的构造器org.apache.spark.mllib.linalg.distributed.CoordinateMatrix
-
- CoordinateMatrix(RDD<MatrixEntry>) - 类 的构造器org.apache.spark.mllib.linalg.distributed.CoordinateMatrix
-
Alternative constructor leaving matrix dimensions to be determined automatically.
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.classification.GBTClassificationModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.classification.GBTClassifier
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.classification.LinearSVC
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.classification.LinearSVCModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.classification.LogisticRegression
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.classification.NaiveBayes
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.classification.NaiveBayesModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.classification.OneVsRest
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.classification.OneVsRestModel
-
- copy(ParamMap) - 类 中的静态方法org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.classification.RandomForestClassifier
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.clustering.BisectingKMeans
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.clustering.GaussianMixture
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.clustering.KMeans
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.clustering.KMeansModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.clustering.LDA
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.clustering.LocalLDAModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.Estimator
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.evaluation.ClusteringEvaluator
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.evaluation.Evaluator
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.evaluation.RegressionEvaluator
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.Binarizer
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.Bucketizer
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.ChiSqSelector
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.ColumnPruner
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.CountVectorizer
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.CountVectorizerModel
-
- copy(ParamMap) - 类 中的静态方法org.apache.spark.ml.feature.DCT
-
- copy(ParamMap) - 类 中的静态方法org.apache.spark.ml.feature.ElementwiseProduct
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.FeatureHasher
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.HashingTF
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.IDF
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.IDFModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.Imputer
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.ImputerModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.IndexToString
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.Interaction
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.MaxAbsScaler
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.MaxAbsScalerModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.MinHashLSH
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.MinHashLSHModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.MinMaxScaler
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.MinMaxScalerModel
-
- copy(ParamMap) - 类 中的静态方法org.apache.spark.ml.feature.NGram
-
- copy(ParamMap) - 类 中的静态方法org.apache.spark.ml.feature.Normalizer
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.OneHotEncoder
-
已过时。
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.OneHotEncoderEstimator
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.OneHotEncoderModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.PCA
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.PCAModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.PolynomialExpansion
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.QuantileDiscretizer
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.RegexTokenizer
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.RFormula
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.RFormulaModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.SQLTransformer
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.StandardScaler
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.StandardScalerModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.StopWordsRemover
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.StringIndexer
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.StringIndexerModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.Tokenizer
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.VectorAssembler
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.VectorAttributeRewriter
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.VectorIndexer
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.VectorIndexerModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.VectorSizeHint
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.VectorSlicer
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.Word2Vec
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.feature.Word2VecModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.fpm.FPGrowth
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.fpm.FPGrowthModel
-
- copy(Vector, Vector) - 类 中的静态方法org.apache.spark.ml.linalg.BLAS
-
y = x
- copy() - 类 中的方法org.apache.spark.ml.linalg.DenseMatrix
-
- copy() - 类 中的方法org.apache.spark.ml.linalg.DenseVector
-
- copy() - 接口 中的方法org.apache.spark.ml.linalg.Matrix
-
Get a deep copy of the matrix.
- copy() - 类 中的方法org.apache.spark.ml.linalg.SparseMatrix
-
- copy() - 类 中的方法org.apache.spark.ml.linalg.SparseVector
-
- copy() - 接口 中的方法org.apache.spark.ml.linalg.Vector
-
Makes a deep copy of this vector.
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.Model
-
- copy() - 类 中的方法org.apache.spark.ml.param.ParamMap
-
Creates a copy of this param map.
- copy(ParamMap) - 接口 中的方法org.apache.spark.ml.param.Params
-
Creates a copy of this instance with the same UID and some extra params.
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.Pipeline
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.PipelineModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.PipelineStage
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.Predictor
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.recommendation.ALS
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.recommendation.ALSModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.regression.GBTRegressionModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.regression.GBTRegressor
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.regression.IsotonicRegression
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.regression.IsotonicRegressionModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.regression.LinearRegression
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.regression.LinearRegressionModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.regression.RandomForestRegressor
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.Transformer
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.tuning.CrossValidator
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.tuning.CrossValidatorModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.tuning.TrainValidationSplit
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.tuning.TrainValidationSplitModel
-
- copy(ParamMap) - 类 中的方法org.apache.spark.ml.UnaryTransformer
-
- copy(Vector, Vector) - 类 中的静态方法org.apache.spark.mllib.linalg.BLAS
-
y = x
- copy() - 类 中的方法org.apache.spark.mllib.linalg.DenseMatrix
-
- copy() - 类 中的方法org.apache.spark.mllib.linalg.DenseVector
-
- copy() - 接口 中的方法org.apache.spark.mllib.linalg.Matrix
-
Get a deep copy of the matrix.
- copy() - 类 中的方法org.apache.spark.mllib.linalg.SparseMatrix
-
- copy() - 类 中的方法org.apache.spark.mllib.linalg.SparseVector
-
- copy() - 接口 中的方法org.apache.spark.mllib.linalg.Vector
-
Makes a deep copy of this vector.
- copy() - 类 中的方法org.apache.spark.mllib.random.ExponentialGenerator
-
- copy() - 类 中的方法org.apache.spark.mllib.random.GammaGenerator
-
- copy() - 类 中的方法org.apache.spark.mllib.random.LogNormalGenerator
-
- copy() - 类 中的方法org.apache.spark.mllib.random.PoissonGenerator
-
- copy() - 接口 中的方法org.apache.spark.mllib.random.RandomDataGenerator
-
Returns a copy of the RandomDataGenerator with a new instance of the rng object used in the
class when applicable for non-locking concurrent usage.
- copy() - 类 中的方法org.apache.spark.mllib.random.StandardNormalGenerator
-
- copy() - 类 中的方法org.apache.spark.mllib.random.UniformGenerator
-
- copy() - 类 中的方法org.apache.spark.mllib.random.WeibullGenerator
-
- copy() - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
Returns a shallow copy of this instance.
- copy(Kryo, T) - 类 中的静态方法org.apache.spark.serializer.JavaIterableWrapperSerializer
-
- copy() - 接口 中的方法org.apache.spark.sql.Row
-
Make a copy of the current
Row object.
- copy() - 类 中的方法org.apache.spark.sql.vectorized.ColumnarArray
-
- copy() - 类 中的方法org.apache.spark.sql.vectorized.ColumnarMap
-
- copy() - 类 中的方法org.apache.spark.sql.vectorized.ColumnarRow
-
Revisit this.
- copy() - 类 中的方法org.apache.spark.util.AccumulatorV2
-
Creates a new copy of this accumulator.
- copy() - 类 中的方法org.apache.spark.util.CollectionAccumulator
-
- copy() - 类 中的方法org.apache.spark.util.DoubleAccumulator
-
- copy() - 类 中的方法org.apache.spark.util.LegacyAccumulatorWrapper
-
- copy() - 类 中的方法org.apache.spark.util.LongAccumulator
-
- copy() - 类 中的方法org.apache.spark.util.StatCounter
-
Clone this StatCounter
- copyAndReset() - 类 中的方法org.apache.spark.util.AccumulatorV2
-
Creates a new copy of this accumulator, which is zero value. i.e. call isZero on the copy
must return true.
- copyAndReset() - 类 中的方法org.apache.spark.util.CollectionAccumulator
-
- copyFileStreamNIO(FileChannel, FileChannel, long, long) - 类 中的静态方法org.apache.spark.util.Utils
-
- copyStream(InputStream, OutputStream, boolean, boolean) - 类 中的静态方法org.apache.spark.util.Utils
-
Copy all data from an InputStream to an OutputStream.
- copyToArray(Object, int) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- copyToArray(Object) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- copyToArray(Object, int, int) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- copyToBuffer(Buffer<B>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- copyValues(T, ParamMap) - 接口 中的方法org.apache.spark.ml.param.Params
-
Copies param values from this instance to another instance for params shared by them.
- cores() - 类 中的方法org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RegisterExecutor
-
- coresGranted() - 类 中的方法org.apache.spark.status.api.v1.ApplicationInfo
-
- coresPerExecutor() - 类 中的方法org.apache.spark.status.api.v1.ApplicationInfo
-
- corr(Dataset<?>, String, String) - 类 中的静态方法org.apache.spark.ml.stat.Correlation
-
:: Experimental ::
Compute the correlation matrix for the input Dataset of Vectors using the specified method.
- corr(Dataset<?>, String) - 类 中的静态方法org.apache.spark.ml.stat.Correlation
-
Compute the Pearson correlation matrix for the input Dataset of Vectors.
- corr(RDD<Object>, RDD<Object>, String) - 类 中的静态方法org.apache.spark.mllib.stat.correlation.Correlations
-
- corr(RDD<Vector>) - 类 中的静态方法org.apache.spark.mllib.stat.Statistics
-
Compute the Pearson correlation matrix for the input RDD of Vectors.
- corr(RDD<Vector>, String) - 类 中的静态方法org.apache.spark.mllib.stat.Statistics
-
Compute the correlation matrix for the input RDD of Vectors using the specified method.
- corr(RDD<Object>, RDD<Object>) - 类 中的静态方法org.apache.spark.mllib.stat.Statistics
-
Compute the Pearson correlation for the input RDDs.
- corr(JavaRDD<Double>, JavaRDD<Double>) - 类 中的静态方法org.apache.spark.mllib.stat.Statistics
-
Java-friendly version of corr()
- corr(RDD<Object>, RDD<Object>, String) - 类 中的静态方法org.apache.spark.mllib.stat.Statistics
-
Compute the correlation for the input RDDs using the specified method.
- corr(JavaRDD<Double>, JavaRDD<Double>, String) - 类 中的静态方法org.apache.spark.mllib.stat.Statistics
-
Java-friendly version of corr()
- corr(String, String, String) - 类 中的方法org.apache.spark.sql.DataFrameStatFunctions
-
Calculates the correlation of two columns of a DataFrame.
- corr(String, String) - 类 中的方法org.apache.spark.sql.DataFrameStatFunctions
-
Calculates the Pearson Correlation Coefficient of two columns of a DataFrame.
- corr(Column, Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the Pearson Correlation Coefficient for two columns.
- corr(String, String) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the Pearson Correlation Coefficient for two columns.
- Correlation - org.apache.spark.ml.stat中的类
-
API for correlation functions in MLlib, compatible with DataFrames and Datasets.
- Correlation() - 类 的构造器org.apache.spark.ml.stat.Correlation
-
- CorrelationNames - org.apache.spark.mllib.stat.correlation中的类
-
Maintains supported and default correlation names.
- CorrelationNames() - 类 的构造器org.apache.spark.mllib.stat.correlation.CorrelationNames
-
- Correlations - org.apache.spark.mllib.stat.correlation中的类
-
Delegates computation to the specific correlation object based on the input method name.
- Correlations() - 类 的构造器org.apache.spark.mllib.stat.correlation.Correlations
-
- corresponds(GenSeq<B>, Function2<A, B, Object>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- corrMatrix(RDD<Vector>, String) - 类 中的静态方法org.apache.spark.mllib.stat.correlation.Correlations
-
- cos(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
- cos(String) - 类 中的静态方法org.apache.spark.sql.functions
-
- cosh(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
- cosh(String) - 类 中的静态方法org.apache.spark.sql.functions
-
- count() - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- count() - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- count() - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- count() - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Return the number of elements in the RDD.
- count() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- count() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- count() - 类 中的方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
The number of edges in the RDD.
- count() - 类 中的方法org.apache.spark.graphx.impl.VertexRDDImpl
-
The number of vertices in the RDD.
- count() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- count() - 类 中的方法org.apache.spark.ml.clustering.ExpectationAggregator
-
- count() - 类 中的方法org.apache.spark.ml.regression.AFTAggregator
-
- count(Column, Column) - 类 中的静态方法org.apache.spark.ml.stat.Summarizer
-
- count(Column) - 类 中的静态方法org.apache.spark.ml.stat.Summarizer
-
- count() - 类 中的方法org.apache.spark.mllib.stat.MultivariateOnlineSummarizer
-
Sample size.
- count() - 接口 中的方法org.apache.spark.mllib.stat.MultivariateStatisticalSummary
-
Sample size.
- count() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- count() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- count() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- count() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- count() - 类 中的方法org.apache.spark.rdd.RDD
-
Return the number of elements in the RDD.
- count() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- count() - 类 中的方法org.apache.spark.sql.Dataset
-
Returns the number of rows in the Dataset.
- count(MapFunction<T, Object>) - 类 中的静态方法org.apache.spark.sql.expressions.javalang.typed
-
Count aggregate function.
- count(Function1<IN, Object>) - 类 中的静态方法org.apache.spark.sql.expressions.scalalang.typed
-
Count aggregate function.
- count(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the number of items in a group.
- count(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the number of items in a group.
- count() - 类 中的方法org.apache.spark.sql.KeyValueGroupedDataset
-
Returns a
Dataset that contains a tuple with each key and the number of items present
for that key.
- count() - 类 中的方法org.apache.spark.sql.RelationalGroupedDataset
-
Count the number of rows for each group.
- count(Function1<A, Object>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- count() - 类 中的方法org.apache.spark.status.RDDPartitionSeq
-
- count() - 类 中的方法org.apache.spark.storage.ReadableChannelFileRegion
-
- count() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- count() - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return a new DStream in which each RDD has a single element generated by counting each RDD
of this DStream.
- count() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- count() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- count() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- count() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- count() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- count() - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Return a new DStream in which each RDD has a single element generated by counting each RDD
of this DStream.
- count() - 类 中的方法org.apache.spark.util.DoubleAccumulator
-
Returns the number of elements added to the accumulator.
- count(Class<?>) - 类 中的方法org.apache.spark.util.kvstore.InMemoryStore
-
- count(Class<?>, String, Object) - 类 中的方法org.apache.spark.util.kvstore.InMemoryStore
-
- count(Class<?>) - 接口 中的方法org.apache.spark.util.kvstore.KVStore
-
Returns the number of items of the given type currently in the store.
- count(Class<?>, String, Object) - 接口 中的方法org.apache.spark.util.kvstore.KVStore
-
Returns the number of items of the given type which match the given indexed value.
- count(Class<?>) - 类 中的方法org.apache.spark.util.kvstore.LevelDB
-
- count(Class<?>, String, Object) - 类 中的方法org.apache.spark.util.kvstore.LevelDB
-
- count() - 类 中的方法org.apache.spark.util.LongAccumulator
-
Returns the number of elements added to the accumulator.
- count() - 类 中的方法org.apache.spark.util.StatCounter
-
- countApprox(long, double) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- countApprox(long) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- countApprox(long, double) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- countApprox(long) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- countApprox(long, double) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- countApprox(long) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- countApprox(long, double) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Approximate version of count() that returns a potentially incomplete result
within a timeout, even if not all tasks have finished.
- countApprox(long) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Approximate version of count() that returns a potentially incomplete result
within a timeout, even if not all tasks have finished.
- countApprox(long, double) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- countApprox(long, double) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- countApprox(long, double) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- countApprox(long, double) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- countApprox(long, double) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- countApprox(long, double) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- countApprox(long, double) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- countApprox(long, double) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- countApprox(long, double) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- countApprox(long, double) - 类 中的方法org.apache.spark.rdd.RDD
-
Approximate version of count() that returns a potentially incomplete result
within a timeout, even if not all tasks have finished.
- countApprox(long, double) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- countApprox$default$2() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- countApprox$default$2() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- countApprox$default$2() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- countApprox$default$2() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- countApprox$default$2() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- countApprox$default$2() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- countApprox$default$2() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- countApprox$default$2() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- countApprox$default$2() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- countApprox$default$2() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- countApproxDistinct(double) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- countApproxDistinct(double) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- countApproxDistinct(double) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- countApproxDistinct(double) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Return approximate number of distinct elements in the RDD.
- countApproxDistinct(int, int) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- countApproxDistinct(double) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- countApproxDistinct(int, int) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- countApproxDistinct(double) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- countApproxDistinct(int, int) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- countApproxDistinct(double) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- countApproxDistinct(int, int) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- countApproxDistinct(double) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- countApproxDistinct(int, int) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- countApproxDistinct(double) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- countApproxDistinct(int, int) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- countApproxDistinct(double) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- countApproxDistinct(int, int) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- countApproxDistinct(double) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- countApproxDistinct(int, int) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- countApproxDistinct(double) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- countApproxDistinct(int, int) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- countApproxDistinct(double) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- countApproxDistinct(int, int) - 类 中的方法org.apache.spark.rdd.RDD
-
Return approximate number of distinct elements in the RDD.
- countApproxDistinct(double) - 类 中的方法org.apache.spark.rdd.RDD
-
Return approximate number of distinct elements in the RDD.
- countApproxDistinct(int, int) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- countApproxDistinct(double) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- countApproxDistinct$default$1() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- countApproxDistinct$default$1() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- countApproxDistinct$default$1() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- countApproxDistinct$default$1() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- countApproxDistinct$default$1() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- countApproxDistinct$default$1() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- countApproxDistinct$default$1() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- countApproxDistinct$default$1() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- countApproxDistinct$default$1() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- countApproxDistinct$default$1() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- countApproxDistinctByKey(double, Partitioner) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Return approximate number of distinct values for each key in this RDD.
- countApproxDistinctByKey(double, int) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Return approximate number of distinct values for each key in this RDD.
- countApproxDistinctByKey(double) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Return approximate number of distinct values for each key in this RDD.
- countApproxDistinctByKey(int, int, Partitioner) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Return approximate number of distinct values for each key in this RDD.
- countApproxDistinctByKey(double, Partitioner) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Return approximate number of distinct values for each key in this RDD.
- countApproxDistinctByKey(double, int) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Return approximate number of distinct values for each key in this RDD.
- countApproxDistinctByKey(double) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Return approximate number of distinct values for each key in this RDD.
- countAsync() - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- countAsync() - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- countAsync() - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- countAsync() - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
The asynchronous version of count, which returns a
future for counting the number of elements in this RDD.
- countAsync() - 类 中的方法org.apache.spark.rdd.AsyncRDDActions
-
Returns a future for counting the number of elements in the RDD.
- countByKey() - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Count the number of elements for each key, and return the result to the master as a Map.
- countByKey() - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Count the number of elements for each key, collecting the results to a local Map.
- countByKeyApprox(long) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Approximate version of countByKey that can return a partial result if it does
not finish within a timeout.
- countByKeyApprox(long, double) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Approximate version of countByKey that can return a partial result if it does
not finish within a timeout.
- countByKeyApprox(long, double) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Approximate version of countByKey that can return a partial result if it does
not finish within a timeout.
- countByValue() - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- countByValue() - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- countByValue() - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- countByValue() - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Return the count of each unique value in this RDD as a map of (value, count) pairs.
- countByValue(Ordering<T>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- countByValue(Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- countByValue(Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- countByValue(Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- countByValue(Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- countByValue(Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- countByValue(Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- countByValue(Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- countByValue(Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- countByValue(Ordering<T>) - 类 中的方法org.apache.spark.rdd.RDD
-
Return the count of each unique value in this RDD as a local map of (value, count) pairs.
- countByValue(Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- countByValue() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- countByValue(int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- countByValue() - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return a new DStream in which each RDD contains the counts of each distinct value in
each RDD of this DStream.
- countByValue(int) - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return a new DStream in which each RDD contains the counts of each distinct value in
each RDD of this DStream.
- countByValue() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- countByValue(int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- countByValue() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- countByValue(int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- countByValue() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- countByValue(int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- countByValue() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- countByValue(int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- countByValue() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- countByValue(int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- countByValue(int, Ordering<T>) - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Return a new DStream in which each RDD contains the counts of each distinct value in
each RDD of this DStream.
- countByValue$default$1() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- countByValue$default$1() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- countByValue$default$1() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- countByValue$default$1() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- countByValue$default$1() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- countByValue$default$1() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- countByValue$default$1() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- countByValue$default$1() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- countByValue$default$1() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- countByValue$default$1() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- countByValueAndWindow(Duration, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- countByValueAndWindow(Duration, Duration, int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- countByValueAndWindow(Duration, Duration) - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return a new DStream in which each RDD contains the count of distinct elements in
RDDs in a sliding window over this DStream.
- countByValueAndWindow(Duration, Duration, int) - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return a new DStream in which each RDD contains the count of distinct elements in
RDDs in a sliding window over this DStream.
- countByValueAndWindow(Duration, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- countByValueAndWindow(Duration, Duration, int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- countByValueAndWindow(Duration, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- countByValueAndWindow(Duration, Duration, int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- countByValueAndWindow(Duration, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- countByValueAndWindow(Duration, Duration, int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- countByValueAndWindow(Duration, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- countByValueAndWindow(Duration, Duration, int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- countByValueAndWindow(Duration, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- countByValueAndWindow(Duration, Duration, int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- countByValueAndWindow(Duration, Duration, int, Ordering<T>) - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Return a new DStream in which each RDD contains the count of distinct elements in
RDDs in a sliding window over this DStream.
- countByValueApprox(long, double) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- countByValueApprox(long) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- countByValueApprox(long, double) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- countByValueApprox(long) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- countByValueApprox(long, double) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- countByValueApprox(long) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- countByValueApprox(long, double) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Approximate version of countByValue().
- countByValueApprox(long) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Approximate version of countByValue().
- countByValueApprox(long, double, Ordering<T>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- countByValueApprox(long, double, Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- countByValueApprox(long, double, Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- countByValueApprox(long, double, Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- countByValueApprox(long, double, Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- countByValueApprox(long, double, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- countByValueApprox(long, double, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- countByValueApprox(long, double, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- countByValueApprox(long, double, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- countByValueApprox(long, double, Ordering<T>) - 类 中的方法org.apache.spark.rdd.RDD
-
Approximate version of countByValue().
- countByValueApprox(long, double, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- countByValueApprox$default$2() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- countByValueApprox$default$2() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- countByValueApprox$default$2() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- countByValueApprox$default$2() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- countByValueApprox$default$2() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- countByValueApprox$default$2() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- countByValueApprox$default$2() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- countByValueApprox$default$2() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- countByValueApprox$default$2() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- countByValueApprox$default$2() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- countByValueApprox$default$3(long, double) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- countByValueApprox$default$3(long, double) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- countByValueApprox$default$3(long, double) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- countByValueApprox$default$3(long, double) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- countByValueApprox$default$3(long, double) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- countByValueApprox$default$3(long, double) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- countByValueApprox$default$3(long, double) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- countByValueApprox$default$3(long, double) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- countByValueApprox$default$3(long, double) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- countByValueApprox$default$3(long, double) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- countByWindow(Duration, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- countByWindow(Duration, Duration) - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return a new DStream in which each RDD has a single element generated by counting the number
of elements in a window over this DStream. windowDuration and slideDuration are as defined in
the window() operation.
- countByWindow(Duration, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- countByWindow(Duration, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- countByWindow(Duration, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- countByWindow(Duration, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- countByWindow(Duration, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- countByWindow(Duration, Duration) - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Return a new DStream in which each RDD has a single element generated by counting the number
of elements in a sliding window over this DStream.
- countDistinct(Column, Column...) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the number of distinct items in a group.
- countDistinct(String, String...) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the number of distinct items in a group.
- countDistinct(Column, Seq<Column>) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the number of distinct items in a group.
- countDistinct(String, Seq<String>) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the number of distinct items in a group.
- COUNTER() - 类 中的静态方法org.apache.spark.metrics.sink.StatsdMetricType
-
- CountingWritableChannel - org.apache.spark.storage中的类
-
- CountingWritableChannel(WritableByteChannel) - 类 的构造器org.apache.spark.storage.CountingWritableChannel
-
- countMinSketch(String, int, int, int) - 类 中的方法org.apache.spark.sql.DataFrameStatFunctions
-
Builds a Count-min Sketch over a specified column.
- countMinSketch(String, double, double, int) - 类 中的方法org.apache.spark.sql.DataFrameStatFunctions
-
Builds a Count-min Sketch over a specified column.
- countMinSketch(Column, int, int, int) - 类 中的方法org.apache.spark.sql.DataFrameStatFunctions
-
Builds a Count-min Sketch over a specified column.
- countMinSketch(Column, double, double, int) - 类 中的方法org.apache.spark.sql.DataFrameStatFunctions
-
Builds a Count-min Sketch over a specified column.
- CountMinSketch - org.apache.spark.util.sketch中的类
-
A Count-min sketch is a probabilistic data structure used for cardinality estimation using
sub-linear space.
- CountMinSketch() - 类 的构造器org.apache.spark.util.sketch.CountMinSketch
-
- CountMinSketch.Version - org.apache.spark.util.sketch中的枚举
-
- countTowardsTaskFailures() - 类 中的静态方法org.apache.spark.ExceptionFailure
-
- countTowardsTaskFailures() - 类 中的方法org.apache.spark.ExecutorLostFailure
-
- countTowardsTaskFailures() - 类 中的方法org.apache.spark.FetchFailed
-
Fetch failures lead to a different failure handling path: (1) we don't abort the stage after
4 task failures, instead we immediately go back to the stage which generated the map output,
and regenerate the missing data
- countTowardsTaskFailures() - 类 中的静态方法org.apache.spark.Resubmitted
-
- countTowardsTaskFailures() - 类 中的方法org.apache.spark.TaskCommitDenied
-
If a task failed because its attempt to commit was denied, do not count this failure
towards failing the stage.
- countTowardsTaskFailures() - 接口 中的方法org.apache.spark.TaskFailedReason
-
Whether this task failure should be counted towards the maximum number of times the task is
allowed to fail before the stage is aborted.
- countTowardsTaskFailures() - 类 中的方法org.apache.spark.TaskKilled
-
- countTowardsTaskFailures() - 类 中的静态方法org.apache.spark.TaskResultLost
-
- countTowardsTaskFailures() - 类 中的静态方法org.apache.spark.UnknownReason
-
- CountVectorizer - org.apache.spark.ml.feature中的类
-
- CountVectorizer(String) - 类 的构造器org.apache.spark.ml.feature.CountVectorizer
-
- CountVectorizer() - 类 的构造器org.apache.spark.ml.feature.CountVectorizer
-
- CountVectorizerModel - org.apache.spark.ml.feature中的类
-
Converts a text document to a sparse vector of token counts.
- CountVectorizerModel(String, String[]) - 类 的构造器org.apache.spark.ml.feature.CountVectorizerModel
-
- CountVectorizerModel(String[]) - 类 的构造器org.apache.spark.ml.feature.CountVectorizerModel
-
- cov() - 类 中的方法org.apache.spark.ml.stat.distribution.MultivariateGaussian
-
- cov(String, String) - 类 中的方法org.apache.spark.sql.DataFrameStatFunctions
-
Calculate the sample covariance of two numerical columns of a DataFrame.
- covar_pop(Column, Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the population covariance for two columns.
- covar_pop(String, String) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the population covariance for two columns.
- covar_samp(Column, Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the sample covariance for two columns.
- covar_samp(String, String) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the sample covariance for two columns.
- covs() - 类 中的方法org.apache.spark.ml.clustering.ExpectationAggregator
-
- crc32(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Calculates the cyclic redundancy check value (CRC32) of a binary column and
returns the value as a bigint.
- CreatableRelationProvider - org.apache.spark.sql.sources中的接口
-
- create(boolean, boolean, boolean, boolean, int) - 类 中的静态方法org.apache.spark.api.java.StorageLevels
-
Create a new StorageLevel object.
- create(JavaSparkContext, JdbcRDD.ConnectionFactory, String, long, long, int, Function<ResultSet, T>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
Create an RDD that executes a SQL query on a JDBC connection and reads results.
- create(JavaSparkContext, JdbcRDD.ConnectionFactory, String, long, long, int) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
Create an RDD that executes a SQL query on a JDBC connection and reads results.
- create(RDD<T>, Function1<Object, Object>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
Create a PartitionPruningRDD.
- create(Object...) - 类 中的静态方法org.apache.spark.sql.RowFactory
-
Create a
Row from the given arguments.
- create(String) - 类 中的静态方法org.apache.spark.sql.streaming.ProcessingTime
-
- create(long, TimeUnit) - 类 中的静态方法org.apache.spark.sql.streaming.ProcessingTime
-
- create(long) - 类 中的静态方法org.apache.spark.util.sketch.BloomFilter
-
Creates a
BloomFilter with the expected number of insertions and a default expected
false positive probability of 3%.
- create(long, double) - 类 中的静态方法org.apache.spark.util.sketch.BloomFilter
-
Creates a
BloomFilter with the expected number of insertions and expected false
positive probability.
- create(long, long) - 类 中的静态方法org.apache.spark.util.sketch.BloomFilter
-
Creates a
BloomFilter with given
expectedNumItems and
numBits, it will
pick an optimal
numHashFunctions which can minimize
fpp for the bloom filter.
- create(int, int, int) - 类 中的静态方法org.apache.spark.util.sketch.CountMinSketch
-
- create(double, double, int) - 类 中的静态方法org.apache.spark.util.sketch.CountMinSketch
-
Creates a
CountMinSketch with given relative error (
eps),
confidence,
and random
seed.
- createArrayType(DataType) - 类 中的静态方法org.apache.spark.sql.types.DataTypes
-
Creates an ArrayType by specifying the data type of elements (elementType).
- createArrayType(DataType, boolean) - 类 中的静态方法org.apache.spark.sql.types.DataTypes
-
Creates an ArrayType by specifying the data type of elements (elementType) and
whether the array contains null values (containsNull).
- createAttrGroupForAttrNames(String, int, boolean, boolean) - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderCommon
-
Creates an `AttributeGroup` with the required number of `BinaryAttribute`.
- createBatchDataReaderFactories() - 接口 中的方法org.apache.spark.sql.sources.v2.reader.SupportsScanColumnarBatch
-
- createCombiner() - 类 中的方法org.apache.spark.Aggregator
-
- createCommitter(int) - 类 中的方法org.apache.spark.internal.io.HadoopWriteConfigUtil
-
- createCompiledClass(String, File, TestUtils.JavaSourceFromString, Seq<URL>) - 类 中的静态方法org.apache.spark.TestUtils
-
Creates a compiled class with the source file.
- createCompiledClass(String, File, String, String, Seq<URL>) - 类 中的静态方法org.apache.spark.TestUtils
-
Creates a compiled class with the given name.
- createContinuousReader(Optional<StructType>, String, DataSourceOptions) - 接口 中的方法org.apache.spark.sql.sources.v2.ContinuousReadSupport
-
- createCryptoInputStream(InputStream, SparkConf, byte[]) - 类 中的静态方法org.apache.spark.security.CryptoStreamUtils
-
Helper method to wrap InputStream with CryptoInputStream for decryption.
- createCryptoOutputStream(OutputStream, SparkConf, byte[]) - 类 中的静态方法org.apache.spark.security.CryptoStreamUtils
-
Helper method to wrap OutputStream with CryptoOutputStream for encryption.
- createDataFrame(RDD<A>, TypeTags.TypeTag<A>) - 类 中的方法org.apache.spark.sql.SparkSession
-
:: Experimental ::
Creates a DataFrame from an RDD of Product (e.g. case classes, tuples).
- createDataFrame(Seq<A>, TypeTags.TypeTag<A>) - 类 中的方法org.apache.spark.sql.SparkSession
-
:: Experimental ::
Creates a DataFrame from a local Seq of Product.
- createDataFrame(RDD<Row>, StructType) - 类 中的方法org.apache.spark.sql.SparkSession
-
:: DeveloperApi ::
Creates a
DataFrame from an
RDD containing
Rows using the given schema.
- createDataFrame(JavaRDD<Row>, StructType) - 类 中的方法org.apache.spark.sql.SparkSession
-
:: DeveloperApi ::
Creates a
DataFrame from a
JavaRDD containing
Rows using the given schema.
- createDataFrame(List<Row>, StructType) - 类 中的方法org.apache.spark.sql.SparkSession
-
:: DeveloperApi ::
Creates a
DataFrame from a
java.util.List containing
Rows using the given schema.
- createDataFrame(RDD<?>, Class<?>) - 类 中的方法org.apache.spark.sql.SparkSession
-
Applies a schema to an RDD of Java Beans.
- createDataFrame(JavaRDD<?>, Class<?>) - 类 中的方法org.apache.spark.sql.SparkSession
-
Applies a schema to an RDD of Java Beans.
- createDataFrame(List<?>, Class<?>) - 类 中的方法org.apache.spark.sql.SparkSession
-
Applies a schema to a List of Java Beans.
- createDataFrame(RDD<A>, TypeTags.TypeTag<A>) - 类 中的方法org.apache.spark.sql.SQLContext
-
- createDataFrame(Seq<A>, TypeTags.TypeTag<A>) - 类 中的方法org.apache.spark.sql.SQLContext
-
- createDataFrame(RDD<Row>, StructType) - 类 中的方法org.apache.spark.sql.SQLContext
-
- createDataFrame(JavaRDD<Row>, StructType) - 类 中的方法org.apache.spark.sql.SQLContext
-
- createDataFrame(List<Row>, StructType) - 类 中的方法org.apache.spark.sql.SQLContext
-
- createDataFrame(RDD<?>, Class<?>) - 类 中的方法org.apache.spark.sql.SQLContext
-
- createDataFrame(JavaRDD<?>, Class<?>) - 类 中的方法org.apache.spark.sql.SQLContext
-
- createDataFrame(List<?>, Class<?>) - 类 中的方法org.apache.spark.sql.SQLContext
-
- createDataReader() - 接口 中的方法org.apache.spark.sql.sources.v2.reader.DataReaderFactory
-
Returns a data reader to do the actual reading work.
- createDataReaderFactories() - 接口 中的方法org.apache.spark.sql.sources.v2.reader.DataSourceReader
-
Returns a list of reader factories.
- createDataReaderFactories() - 接口 中的方法org.apache.spark.sql.sources.v2.reader.SupportsScanColumnarBatch
-
- createDataReaderFactories() - 接口 中的方法org.apache.spark.sql.sources.v2.reader.SupportsScanUnsafeRow
-
- createDataset(Seq<T>, Encoder<T>) - 类 中的方法org.apache.spark.sql.SparkSession
-
:: Experimental ::
Creates a
Dataset from a local Seq of data of a given type.
- createDataset(RDD<T>, Encoder<T>) - 类 中的方法org.apache.spark.sql.SparkSession
-
:: Experimental ::
Creates a
Dataset from an RDD of a given type.
- createDataset(List<T>, Encoder<T>) - 类 中的方法org.apache.spark.sql.SparkSession
-
:: Experimental ::
Creates a
Dataset from a
java.util.List of a given type.
- createDataset(Seq<T>, Encoder<T>) - 类 中的方法org.apache.spark.sql.SQLContext
-
- createDataset(RDD<T>, Encoder<T>) - 类 中的方法org.apache.spark.sql.SQLContext
-
- createDataset(List<T>, Encoder<T>) - 类 中的方法org.apache.spark.sql.SQLContext
-
- createDataWriter(int, int) - 接口 中的方法org.apache.spark.sql.sources.v2.writer.DataWriterFactory
-
Returns a data writer to do the actual writing work.
- createDecimalType(int, int) - 类 中的静态方法org.apache.spark.sql.types.DataTypes
-
Creates a DecimalType by specifying the precision and scale.
- createDecimalType() - 类 中的静态方法org.apache.spark.sql.types.DataTypes
-
Creates a DecimalType with default precision and scale, which are 10 and 0.
- createDF(RDD<byte[]>, StructType, SparkSession) - 类 中的静态方法org.apache.spark.sql.api.r.SQLUtils
-
- createDirectory(String, String) - 类 中的静态方法org.apache.spark.util.Utils
-
Create a directory inside the given parent directory.
- createdTempDir() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- createdTempDir() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- createdTempDir_$eq(Option<Path>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- createdTempDir_$eq(Option<Path>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- createExternalTable(String, String) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
- createExternalTable(String, String, String) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
- createExternalTable(String, String, Map<String, String>) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
- createExternalTable(String, String, Map<String, String>) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
- createExternalTable(String, String, StructType, Map<String, String>) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
- createExternalTable(String, String, StructType, Map<String, String>) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
- createExternalTable(String, String) - 类 中的方法org.apache.spark.sql.SQLContext
-
- createExternalTable(String, String, String) - 类 中的方法org.apache.spark.sql.SQLContext
-
- createExternalTable(String, String, Map<String, String>) - 类 中的方法org.apache.spark.sql.SQLContext
-
- createExternalTable(String, String, Map<String, String>) - 类 中的方法org.apache.spark.sql.SQLContext
-
- createExternalTable(String, String, StructType, Map<String, String>) - 类 中的方法org.apache.spark.sql.SQLContext
-
- createExternalTable(String, String, StructType, Map<String, String>) - 类 中的方法org.apache.spark.sql.SQLContext
-
- createFilter(StructType, Filter[]) - 类 中的静态方法org.apache.spark.sql.hive.orc.OrcFilters
-
- createGlobalTempView(String) - 类 中的方法org.apache.spark.sql.Dataset
-
Creates a global temporary view using the given name.
- CreateHiveTableAsSelectCommand - org.apache.spark.sql.hive.execution中的类
-
Create table and insert the query result into it.
- CreateHiveTableAsSelectCommand(CatalogTable, LogicalPlan, Seq<Attribute>, SaveMode) - 类 的构造器org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- createInternalRowWriterFactory() - 接口 中的方法org.apache.spark.sql.sources.v2.writer.SupportsWriteInternalRow
-
- createJar(Seq<File>, File, Option<String>) - 类 中的静态方法org.apache.spark.TestUtils
-
Create a jar file that contains this set of files.
- createJarWithClasses(Seq<String>, String, Seq<Tuple2<String, String>>, Seq<URL>) - 类 中的静态方法org.apache.spark.TestUtils
-
Create a jar that defines classes with the given names.
- createJarWithFiles(Map<String, String>, File) - 类 中的静态方法org.apache.spark.TestUtils
-
Create a jar file containing multiple files.
- createJobContext(String, int) - 类 中的方法org.apache.spark.internal.io.HadoopWriteConfigUtil
-
- createJobID(Date, int) - 类 中的静态方法org.apache.spark.internal.io.SparkHadoopWriterUtils
-
- createJobTrackerID(Date) - 类 中的静态方法org.apache.spark.internal.io.SparkHadoopWriterUtils
-
- createKey(SparkConf) - 类 中的静态方法org.apache.spark.security.CryptoStreamUtils
-
Creates a new encryption key.
- createLogForDriver(SparkConf, String, Configuration) - 类 中的静态方法org.apache.spark.streaming.util.WriteAheadLogUtils
-
Create a WriteAheadLog for the driver.
- createLogForReceiver(SparkConf, String, Configuration) - 类 中的静态方法org.apache.spark.streaming.util.WriteAheadLogUtils
-
Create a WriteAheadLog for the receiver.
- createMapType(DataType, DataType) - 类 中的静态方法org.apache.spark.sql.types.DataTypes
-
Creates a MapType by specifying the data type of keys (keyType) and values
(keyType).
- createMapType(DataType, DataType, boolean) - 类 中的静态方法org.apache.spark.sql.types.DataTypes
-
Creates a MapType by specifying the data type of keys (keyType), the data type of
values (keyType), and whether values contain any null value
(valueContainsNull).
- createMetrics(long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long, long) - 类 中的静态方法org.apache.spark.status.LiveEntityHelpers
-
- createMetrics(long) - 类 中的静态方法org.apache.spark.status.LiveEntityHelpers
-
- createMicroBatchReader(Optional<StructType>, String, DataSourceOptions) - 接口 中的方法org.apache.spark.sql.sources.v2.MicroBatchReadSupport
-
Creates a
MicroBatchReader to read batches of data from this data source in a
streaming query.
- createOrReplaceGlobalTempView(String) - 类 中的方法org.apache.spark.sql.Dataset
-
Creates or replaces a global temporary view using the given name.
- createOrReplaceTempView(String) - 类 中的方法org.apache.spark.sql.Dataset
-
Creates a local temporary view using the given name.
- createOutputOperationFailureForUI(String) - 类 中的静态方法org.apache.spark.streaming.ui.UIUtils
-
- createPathFromString(String, JobConf) - 类 中的静态方法org.apache.spark.internal.io.SparkHadoopWriterUtils
-
- createPMMLModelExport(Object) - 类 中的静态方法org.apache.spark.mllib.pmml.export.PMMLModelExportFactory
-
Factory object to help creating the necessary PMMLModelExport implementation
taking as input the machine learning model (for example KMeansModel).
- createProxyHandler(Function1<String, Option<String>>) - 类 中的静态方法org.apache.spark.ui.JettyUtils
-
Create a handler for proxying request to Workers and Application Drivers
- createProxyLocationHeader(String, HttpServletRequest, URI) - 类 中的静态方法org.apache.spark.ui.JettyUtils
-
- createProxyURI(String, String, String, String) - 类 中的静态方法org.apache.spark.ui.JettyUtils
-
- createRDDFromArray(JavaSparkContext, byte[][]) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
Create an RRDD given a sequence of byte arrays.
- createRDDFromFile(JavaSparkContext, String, int) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
Create an RRDD given a temporary file name.
- createReadableChannel(ReadableByteChannel, SparkConf, byte[]) - 类 中的静态方法org.apache.spark.security.CryptoStreamUtils
-
Wrap a ReadableByteChannel for decryption.
- createReader(DataSourceOptions) - 接口 中的方法org.apache.spark.sql.sources.v2.ReadSupport
-
- createReader(StructType, DataSourceOptions) - 接口 中的方法org.apache.spark.sql.sources.v2.ReadSupportWithSchema
-
- createRedirectHandler(String, String, Function1<HttpServletRequest, BoxedUnit>, String, Set<String>) - 类 中的静态方法org.apache.spark.ui.JettyUtils
-
Create a handler that always redirects the user to the given path
- createRelation(SQLContext, SaveMode, Map<String, String>, Dataset<Row>) - 接口 中的方法org.apache.spark.sql.sources.CreatableRelationProvider
-
Saves a DataFrame to a destination (using data source-specific parameters)
- createRelation(SQLContext, Map<String, String>) - 接口 中的方法org.apache.spark.sql.sources.RelationProvider
-
Returns a new base relation with the given parameters.
- createRelation(SQLContext, Map<String, String>, StructType) - 接口 中的方法org.apache.spark.sql.sources.SchemaRelationProvider
-
Returns a new base relation with the given parameters and user defined schema.
- createSecret(SparkConf) - 类 中的静态方法org.apache.spark.util.Utils
-
- createServlet(JettyUtils.ServletParams<T>, org.apache.spark.SecurityManager, SparkConf) - 类 中的静态方法org.apache.spark.ui.JettyUtils
-
- createServletHandler(String, JettyUtils.ServletParams<T>, org.apache.spark.SecurityManager, SparkConf, String) - 类 中的静态方法org.apache.spark.ui.JettyUtils
-
Create a context handler that responds to a request with the given path prefix
- createServletHandler(String, HttpServlet, String) - 类 中的静态方法org.apache.spark.ui.JettyUtils
-
Create a context handler that responds to a request with the given path prefix
- createSink(SQLContext, Map<String, String>, Seq<String>, OutputMode) - 接口 中的方法org.apache.spark.sql.sources.StreamSinkProvider
-
- createSource(SQLContext, String, Option<StructType>, String, Map<String, String>) - 接口 中的方法org.apache.spark.sql.sources.StreamSourceProvider
-
- createSparkContext(String, String, String, String[], Map<Object, Object>, Map<Object, Object>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- createStaticHandler(String, String) - 类 中的静态方法org.apache.spark.ui.JettyUtils
-
Create a handler for serving files from a static directory
- createStream(StreamingContext, String, String, String, String, InitialPositionInStream, Duration, StorageLevel, Function1<Record, T>, ClassTag<T>) - 类 中的静态方法org.apache.spark.streaming.kinesis.KinesisUtils
-
- createStream(StreamingContext, String, String, String, String, InitialPositionInStream, Duration, StorageLevel, Function1<Record, T>, String, String, ClassTag<T>) - 类 中的静态方法org.apache.spark.streaming.kinesis.KinesisUtils
-
- createStream(StreamingContext, String, String, String, String, InitialPositionInStream, Duration, StorageLevel, Function1<Record, T>, String, String, String, String, String, ClassTag<T>) - 类 中的静态方法org.apache.spark.streaming.kinesis.KinesisUtils
-
- createStream(StreamingContext, String, String, String, String, InitialPositionInStream, Duration, StorageLevel) - 类 中的静态方法org.apache.spark.streaming.kinesis.KinesisUtils
-
- createStream(StreamingContext, String, String, String, String, InitialPositionInStream, Duration, StorageLevel, String, String) - 类 中的静态方法org.apache.spark.streaming.kinesis.KinesisUtils
-
- createStream(JavaStreamingContext, String, String, String, String, InitialPositionInStream, Duration, StorageLevel, Function<Record, T>, Class<T>) - 类 中的静态方法org.apache.spark.streaming.kinesis.KinesisUtils
-
- createStream(JavaStreamingContext, String, String, String, String, InitialPositionInStream, Duration, StorageLevel, Function<Record, T>, Class<T>, String, String) - 类 中的静态方法org.apache.spark.streaming.kinesis.KinesisUtils
-
- createStream(JavaStreamingContext, String, String, String, String, InitialPositionInStream, Duration, StorageLevel, Function<Record, T>, Class<T>, String, String, String, String, String) - 类 中的静态方法org.apache.spark.streaming.kinesis.KinesisUtils
-
- createStream(JavaStreamingContext, String, String, String, String, InitialPositionInStream, Duration, StorageLevel) - 类 中的静态方法org.apache.spark.streaming.kinesis.KinesisUtils
-
- createStream(JavaStreamingContext, String, String, String, String, InitialPositionInStream, Duration, StorageLevel, String, String) - 类 中的静态方法org.apache.spark.streaming.kinesis.KinesisUtils
-
- createStream(JavaStreamingContext, String, String, String, String, int, Duration, StorageLevel, String, String, String, String, String) - 类 中的方法org.apache.spark.streaming.kinesis.KinesisUtilsPythonHelper
-
- createStreamWriter(String, StructType, OutputMode, DataSourceOptions) - 接口 中的方法org.apache.spark.sql.sources.v2.StreamWriteSupport
-
Creates an optional
StreamWriter to save the data to this data source.
- createStructField(String, String, boolean) - 类 中的静态方法org.apache.spark.sql.api.r.SQLUtils
-
- createStructField(String, DataType, boolean, Metadata) - 类 中的静态方法org.apache.spark.sql.types.DataTypes
-
Creates a StructField by specifying the name (name), data type (dataType) and
whether values of this field can be null values (nullable).
- createStructField(String, DataType, boolean) - 类 中的静态方法org.apache.spark.sql.types.DataTypes
-
Creates a StructField with empty metadata.
- createStructType(Seq<StructField>) - 类 中的静态方法org.apache.spark.sql.api.r.SQLUtils
-
- createStructType(List<StructField>) - 类 中的静态方法org.apache.spark.sql.types.DataTypes
-
Creates a StructType with the given list of StructFields (fields).
- createStructType(StructField[]) - 类 中的静态方法org.apache.spark.sql.types.DataTypes
-
Creates a StructType with the given StructField array (fields).
- createTable(String, String) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
:: Experimental ::
Creates a table from the given path and returns the corresponding DataFrame.
- createTable(String, String, String) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
:: Experimental ::
Creates a table from the given path based on a data source and returns the corresponding
DataFrame.
- createTable(String, String, Map<String, String>) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
:: Experimental ::
Creates a table based on the dataset in a data source and a set of options.
- createTable(String, String, Map<String, String>) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
:: Experimental ::
(Scala-specific)
Creates a table based on the dataset in a data source and a set of options.
- createTable(String, String, StructType, Map<String, String>) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
:: Experimental ::
Create a table based on the dataset in a data source, a schema and a set of options.
- createTable(String, String, StructType, Map<String, String>) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
:: Experimental ::
(Scala-specific)
Create a table based on the dataset in a data source, a schema and a set of options.
- createTaskAttemptContext(String, int, int, int) - 类 中的方法org.apache.spark.internal.io.HadoopWriteConfigUtil
-
- createTempDir(String, String) - 类 中的静态方法org.apache.spark.util.Utils
-
Create a temporary directory inside the given parent directory.
- createTempView(String) - 类 中的方法org.apache.spark.sql.Dataset
-
Creates a local temporary view using the given name.
- createUnsafe(long, int, int) - 类 中的静态方法org.apache.spark.sql.types.Decimal
-
Creates a decimal from unscaled, precision and scale without checking the bounds.
- createUnsafeRowReaderFactories() - 接口 中的方法org.apache.spark.sql.sources.v2.reader.SupportsScanUnsafeRow
-
- createWorkspace(int) - 类 中的静态方法org.apache.spark.mllib.optimization.NNLS
-
- createWritableChannel(WritableByteChannel, SparkConf, byte[]) - 类 中的静态方法org.apache.spark.security.CryptoStreamUtils
-
Wrap a WritableByteChannel for encryption.
- createWriter(String, StructType, SaveMode, DataSourceOptions) - 接口 中的方法org.apache.spark.sql.sources.v2.WriteSupport
-
- createWriterFactory() - 接口 中的方法org.apache.spark.sql.sources.v2.writer.DataSourceWriter
-
Creates a writer factory which will be serialized and sent to executors.
- createWriterFactory() - 接口 中的方法org.apache.spark.sql.sources.v2.writer.SupportsWriteInternalRow
-
- crossJoin(Dataset<?>) - 类 中的方法org.apache.spark.sql.Dataset
-
Explicit cartesian join with another DataFrame.
- crosstab(String, String) - 类 中的方法org.apache.spark.sql.DataFrameStatFunctions
-
Computes a pair-wise frequency table of the given columns.
- CrossValidator - org.apache.spark.ml.tuning中的类
-
K-fold cross validation performs model selection by splitting the dataset into a set of
non-overlapping randomly partitioned folds which are used as separate training and test datasets
e.g., with k=3 folds, K-fold cross validation will generate 3 (training, test) dataset pairs,
each of which uses 2/3 of the data for training and 1/3 for testing.
- CrossValidator(String) - 类 的构造器org.apache.spark.ml.tuning.CrossValidator
-
- CrossValidator() - 类 的构造器org.apache.spark.ml.tuning.CrossValidator
-
- CrossValidatorModel - org.apache.spark.ml.tuning中的类
-
CrossValidatorModel contains the model with the highest average cross-validation
metric across folds and uses this model to transform input data.
- CrossValidatorModel.CrossValidatorModelWriter - org.apache.spark.ml.tuning中的类
-
Writer for CrossValidatorModel.
- CryptoStreamUtils - org.apache.spark.security中的类
-
A util class for manipulating IO encryption and decryption streams.
- CryptoStreamUtils() - 类 的构造器org.apache.spark.security.CryptoStreamUtils
-
- csv(String...) - 类 中的方法org.apache.spark.sql.DataFrameReader
-
Loads CSV files and returns the result as a DataFrame.
- csv(String) - 类 中的方法org.apache.spark.sql.DataFrameReader
-
Loads a CSV file and returns the result as a DataFrame.
- csv(Dataset<String>) - 类 中的方法org.apache.spark.sql.DataFrameReader
-
Loads an Dataset[String] storing CSV rows and returns the result as a DataFrame.
- csv(Seq<String>) - 类 中的方法org.apache.spark.sql.DataFrameReader
-
Loads CSV files and returns the result as a DataFrame.
- csv(String) - 类 中的方法org.apache.spark.sql.DataFrameWriter
-
Saves the content of the DataFrame in CSV format at the specified path.
- csv(String) - 类 中的方法org.apache.spark.sql.streaming.DataStreamReader
-
Loads a CSV file stream and returns the result as a DataFrame.
- cube(Column...) - 类 中的方法org.apache.spark.sql.Dataset
-
Create a multi-dimensional cube for the current Dataset using the specified columns,
so we can run aggregation on them.
- cube(String, String...) - 类 中的方法org.apache.spark.sql.Dataset
-
Create a multi-dimensional cube for the current Dataset using the specified columns,
so we can run aggregation on them.
- cube(Seq<Column>) - 类 中的方法org.apache.spark.sql.Dataset
-
Create a multi-dimensional cube for the current Dataset using the specified columns,
so we can run aggregation on them.
- cube(String, Seq<String>) - 类 中的方法org.apache.spark.sql.Dataset
-
Create a multi-dimensional cube for the current Dataset using the specified columns,
so we can run aggregation on them.
- CubeType$() - 类 的构造器org.apache.spark.sql.RelationalGroupedDataset.CubeType$
-
- cume_dist() - 类 中的静态方法org.apache.spark.sql.functions
-
Window function: returns the cumulative distribution of values within a window partition,
i.e. the fraction of rows that are below the current row.
- current_date() - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the current date as a date column.
- current_timestamp() - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the current timestamp as a timestamp column.
- currentAttemptId() - 接口 中的方法org.apache.spark.SparkStageInfo
-
- currentAttemptId() - 类 中的方法org.apache.spark.SparkStageInfoImpl
-
- currentDatabase() - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
Returns the current default database in this session.
- currentRow() - 类 中的静态方法org.apache.spark.sql.expressions.Window
-
Value representing the current row.
- currentRow() - 类 中的静态方法org.apache.spark.sql.functions
-
Window function: returns the special frame boundary that represents the current row in the
window partition.
- currPrefLocs(Partition, RDD<?>) - 类 中的方法org.apache.spark.rdd.DefaultPartitionCoalescer
-
- dapply(Dataset<Row>, byte[], byte[], Object[], StructType) - 类 中的静态方法org.apache.spark.sql.api.r.SQLUtils
-
The helper function for dapply() on R side.
- Data(Vector, double, Option<Object>) - 类 的构造器org.apache.spark.mllib.classification.impl.GLMClassificationModel.SaveLoadV1_0$.Data
-
- Data(double[], double[], double[][]) - 类 的构造器org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV1_0$.Data
-
- Data(double[], double[], double[][], String) - 类 的构造器org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV2_0$.Data
-
- Data(int) - 类 的构造器org.apache.spark.mllib.feature.ChiSqSelectorModel.SaveLoadV1_0$.Data
-
- Data(Vector, double) - 类 的构造器org.apache.spark.mllib.regression.impl.GLMRegressionModel.SaveLoadV1_0$.Data
-
- data() - 类 中的方法org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.LaunchTask
-
- data() - 类 中的方法org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.StatusUpdate
-
- Database - org.apache.spark.sql.catalog中的类
-
A database in Spark, as returned by the
listDatabases method defined in
Catalog.
- Database(String, String, String) - 类 的构造器org.apache.spark.sql.catalog.Database
-
- database() - 类 中的方法org.apache.spark.sql.catalog.Function
-
- database() - 类 中的方法org.apache.spark.sql.catalog.Table
-
- databaseExists(String) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
Check if the database with the specified name exists.
- databaseTypeDefinition() - 类 中的方法org.apache.spark.sql.jdbc.JdbcType
-
- dataDistribution() - 类 中的方法org.apache.spark.status.api.v1.RDDStorageInfo
-
- DATAFRAME_DAPPLY() - 类 中的静态方法org.apache.spark.api.r.RRunnerModes
-
- DATAFRAME_GAPPLY() - 类 中的静态方法org.apache.spark.api.r.RRunnerModes
-
- DataFrameNaFunctions - org.apache.spark.sql中的类
-
Functionality for working with missing data in DataFrames.
- DataFrameReader - org.apache.spark.sql中的类
-
Interface used to load a
Dataset from external storage systems (e.g. file systems,
key-value stores, etc).
- DataFrameStatFunctions - org.apache.spark.sql中的类
-
Statistic functions for DataFrames.
- DataFrameWriter<T> - org.apache.spark.sql中的类
-
Interface used to write a
Dataset to external storage systems (e.g. file systems,
key-value stores, etc).
- DataReader<T> - org.apache.spark.sql.sources.v2.reader中的接口
-
- DataReaderFactory<T> - org.apache.spark.sql.sources.v2.reader中的接口
-
- Dataset<T> - org.apache.spark.sql中的类
-
A Dataset is a strongly typed collection of domain-specific objects that can be transformed
in parallel using functional or relational operations.
- Dataset(SparkSession, LogicalPlan, Encoder<T>) - 类 的构造器org.apache.spark.sql.Dataset
-
- Dataset(SQLContext, LogicalPlan, Encoder<T>) - 类 的构造器org.apache.spark.sql.Dataset
-
- DatasetHolder<T> - org.apache.spark.sql中的类
-
A container for a
Dataset, used for implicit conversions in Scala.
- DataSourceOptions - org.apache.spark.sql.sources.v2中的类
-
An immutable string-to-string map in which keys are case-insensitive.
- DataSourceOptions(Map<String, String>) - 类 的构造器org.apache.spark.sql.sources.v2.DataSourceOptions
-
- DataSourceReader - org.apache.spark.sql.sources.v2.reader中的接口
-
- DataSourceRegister - org.apache.spark.sql.sources中的接口
-
Data sources should implement this trait so that they can register an alias to their data source.
- DataSourceV2 - org.apache.spark.sql.sources.v2中的接口
-
The base interface for data source v2.
- DataSourceWriter - org.apache.spark.sql.sources.v2.writer中的接口
-
- DataStreamReader - org.apache.spark.sql.streaming中的类
-
Interface used to load a streaming Dataset from external storage systems (e.g. file systems,
key-value stores, etc).
- DataStreamWriter<T> - org.apache.spark.sql.streaming中的类
-
Interface used to write a streaming Dataset to external storage systems (e.g. file systems,
key-value stores, etc).
- dataTablesHeaderNodes() - 类 中的静态方法org.apache.spark.ui.UIUtils
-
- dataType() - 类 中的方法org.apache.spark.sql.catalog.Column
-
- dataType() - 类 中的方法org.apache.spark.sql.expressions.UserDefinedAggregateFunction
-
- dataType() - 类 中的方法org.apache.spark.sql.expressions.UserDefinedFunction
-
- DataType - org.apache.spark.sql.types中的类
-
The base type of all Spark SQL data types.
- DataType() - 类 的构造器org.apache.spark.sql.types.DataType
-
- dataType() - 类 中的方法org.apache.spark.sql.types.StructField
-
- dataType() - 类 中的方法org.apache.spark.sql.vectorized.ColumnVector
-
Returns the data type of this column vector.
- DataTypes - org.apache.spark.sql.types中的类
-
To get/create specific data type, users should use singleton objects and factory methods
provided by this class.
- DataTypes() - 类 的构造器org.apache.spark.sql.types.DataTypes
-
- DataValidators - org.apache.spark.mllib.util中的类
-
:: DeveloperApi ::
A collection of methods used to validate data before applying ML algorithms.
- DataValidators() - 类 的构造器org.apache.spark.mllib.util.DataValidators
-
- DataWriter<T> - org.apache.spark.sql.sources.v2.writer中的接口
-
- DataWriterFactory<T> - org.apache.spark.sql.sources.v2.writer中的接口
-
- date() - 类 中的方法org.apache.spark.sql.ColumnName
-
Creates a new StructField of type date.
- DATE() - 类 中的静态方法org.apache.spark.sql.Encoders
-
An encoder for nullable date type.
- date_add(Column, int) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the date that is days days after start
- date_format(Column, String) - 类 中的静态方法org.apache.spark.sql.functions
-
Converts a date/timestamp/string to a value of string in the format specified by the date
format given by the second argument.
- date_sub(Column, int) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the date that is days days before start
- date_trunc(String, Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns timestamp truncated to the unit specified by the format.
- datediff(Column, Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the number of days from start to end.
- DateType - 类 中的静态变量org.apache.spark.sql.types.DataTypes
-
Gets the DateType object.
- DateType - org.apache.spark.sql.types中的类
-
A date type, supporting "0001-01-01" through "9999-12-31".
- dayofmonth(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Extracts the day of the month as an integer from a given date/timestamp/string.
- dayofweek(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Extracts the day of the week as an integer from a given date/timestamp/string.
- dayofyear(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Extracts the day of the year as an integer from a given date/timestamp/string.
- DB2Dialect - org.apache.spark.sql.jdbc中的类
-
- DB2Dialect() - 类 的构造器org.apache.spark.sql.jdbc.DB2Dialect
-
- DCT - org.apache.spark.ml.feature中的类
-
A feature transformer that takes the 1D discrete cosine transform of a real vector.
- DCT(String) - 类 的构造器org.apache.spark.ml.feature.DCT
-
- DCT() - 类 的构造器org.apache.spark.ml.feature.DCT
-
- deallocate() - 类 中的方法org.apache.spark.storage.ReadableChannelFileRegion
-
- decayFactor() - 类 中的方法org.apache.spark.mllib.clustering.StreamingKMeans
-
- decimal() - 类 中的方法org.apache.spark.sql.ColumnName
-
Creates a new StructField of type decimal.
- decimal(int, int) - 类 中的方法org.apache.spark.sql.ColumnName
-
Creates a new StructField of type decimal.
- DECIMAL() - 类 中的静态方法org.apache.spark.sql.Encoders
-
An encoder for nullable decimal type.
- Decimal - org.apache.spark.sql.types中的类
-
A mutable implementation of BigDecimal that can hold a Long if values are small enough.
- Decimal() - 类 的构造器org.apache.spark.sql.types.Decimal
-
- Decimal.DecimalAsIfIntegral$ - org.apache.spark.sql.types中的类
-
A Integral evidence parameter for Decimals.
- Decimal.DecimalIsFractional$ - org.apache.spark.sql.types中的类
-
A Fractional evidence parameter for Decimals.
- DecimalAsIfIntegral$() - 类 的构造器org.apache.spark.sql.types.Decimal.DecimalAsIfIntegral$
-
- DecimalIsFractional$() - 类 的构造器org.apache.spark.sql.types.Decimal.DecimalIsFractional$
-
- DecimalType - org.apache.spark.sql.types中的类
-
The data type representing java.math.BigDecimal values.
- DecimalType(int, int) - 类 的构造器org.apache.spark.sql.types.DecimalType
-
- DecimalType(int) - 类 的构造器org.apache.spark.sql.types.DecimalType
-
- DecimalType() - 类 的构造器org.apache.spark.sql.types.DecimalType
-
- DecimalType.Expression$ - org.apache.spark.sql.types中的类
-
- DecimalType.Fixed$ - org.apache.spark.sql.types中的类
-
- DecisionTree - org.apache.spark.mllib.tree中的类
-
A class which implements a decision tree learning algorithm for classification and regression.
- DecisionTree(Strategy) - 类 的构造器org.apache.spark.mllib.tree.DecisionTree
-
- DecisionTreeClassificationModel - org.apache.spark.ml.classification中的类
-
Decision tree model (http://en.wikipedia.org/wiki/Decision_tree_learning) for classification.
- DecisionTreeClassifier - org.apache.spark.ml.classification中的类
-
Decision tree learning algorithm (http://en.wikipedia.org/wiki/Decision_tree_learning)
for classification.
- DecisionTreeClassifier(String) - 类 的构造器org.apache.spark.ml.classification.DecisionTreeClassifier
-
- DecisionTreeClassifier() - 类 的构造器org.apache.spark.ml.classification.DecisionTreeClassifier
-
- DecisionTreeModel - org.apache.spark.mllib.tree.model中的类
-
Decision tree model for classification or regression.
- DecisionTreeModel(Node, Enumeration.Value) - 类 的构造器org.apache.spark.mllib.tree.model.DecisionTreeModel
-
- DecisionTreeModel.SaveLoadV1_0$ - org.apache.spark.mllib.tree.model中的类
-
- DecisionTreeModel.SaveLoadV1_0$.NodeData - org.apache.spark.mllib.tree.model中的类
-
Model data for model import/export
- DecisionTreeModel.SaveLoadV1_0$.PredictData - org.apache.spark.mllib.tree.model中的类
-
- DecisionTreeModel.SaveLoadV1_0$.SplitData - org.apache.spark.mllib.tree.model中的类
-
- DecisionTreeModelReadWrite - org.apache.spark.ml.tree中的类
-
Helper classes for tree model persistence
- DecisionTreeModelReadWrite() - 类 的构造器org.apache.spark.ml.tree.DecisionTreeModelReadWrite
-
- DecisionTreeModelReadWrite.NodeData - org.apache.spark.ml.tree中的类
-
Info for a
Node
param: id Index used for tree reconstruction.
- DecisionTreeModelReadWrite.NodeData$ - org.apache.spark.ml.tree中的类
-
- DecisionTreeModelReadWrite.SplitData - org.apache.spark.ml.tree中的类
-
Info for a
Split
param: featureIndex Index of feature split on
param: leftCategoriesOrThreshold For categorical feature, set of leftCategories.
- DecisionTreeModelReadWrite.SplitData$ - org.apache.spark.ml.tree中的类
-
- DecisionTreeRegressionModel - org.apache.spark.ml.regression中的类
-
- DecisionTreeRegressor - org.apache.spark.ml.regression中的类
-
- DecisionTreeRegressor(String) - 类 的构造器org.apache.spark.ml.regression.DecisionTreeRegressor
-
- DecisionTreeRegressor() - 类 的构造器org.apache.spark.ml.regression.DecisionTreeRegressor
-
- decode(Column, String) - 类 中的静态方法org.apache.spark.sql.functions
-
Computes the first argument into a string from a binary using the provided character set
(one of 'US-ASCII', 'ISO-8859-1', 'UTF-8', 'UTF-16BE', 'UTF-16LE', 'UTF-16').
- decodeFileNameInURI(URI) - 类 中的静态方法org.apache.spark.util.Utils
-
Get the file name from uri's raw path and decode it.
- decodeLabel(Vector) - 类 中的静态方法org.apache.spark.ml.classification.LabelConverter
-
Converts a vector to a label.
- decodeURLParameter(String) - 类 中的静态方法org.apache.spark.ui.UIUtils
-
Decode URLParameter if URL is encoded by YARN-WebAppProxyServlet.
- DEFAULT_CONNECTION_TIMEOUT() - 类 中的静态方法org.apache.spark.api.r.SparkRDefaults
-
- DEFAULT_DRIVER_MEM_MB() - 类 中的静态方法org.apache.spark.util.Utils
-
Define a default value for driver memory here since this value is referenced across the code
base and nearly all files already use Utils.scala
- DEFAULT_HEARTBEAT_INTERVAL() - 类 中的静态方法org.apache.spark.api.r.SparkRDefaults
-
- DEFAULT_MAX_FAILURES() - 类 中的静态方法org.apache.spark.streaming.util.WriteAheadLogUtils
-
- DEFAULT_MAX_TO_STRING_FIELDS() - 类 中的静态方法org.apache.spark.util.Utils
-
The performance overhead of creating and logging strings for wide schemas can be large.
- DEFAULT_NUM_RBACKEND_THREADS() - 类 中的静态方法org.apache.spark.api.r.SparkRDefaults
-
- DEFAULT_NUMBER_EXECUTORS() - 类 中的静态方法org.apache.spark.scheduler.cluster.SchedulerBackendUtils
-
- DEFAULT_ROLLING_INTERVAL_SECS() - 类 中的静态方法org.apache.spark.streaming.util.WriteAheadLogUtils
-
- DEFAULT_SHUTDOWN_PRIORITY() - 类 中的静态方法org.apache.spark.util.ShutdownHookManager
-
- defaultAttr() - 类 中的静态方法org.apache.spark.ml.attribute.BinaryAttribute
-
The default binary attribute.
- defaultAttr() - 类 中的静态方法org.apache.spark.ml.attribute.NominalAttribute
-
The default nominal attribute.
- defaultAttr() - 类 中的静态方法org.apache.spark.ml.attribute.NumericAttribute
-
The default numeric attribute.
- defaultCopy(ParamMap) - 接口 中的方法org.apache.spark.ml.param.Params
-
Default implementation of copy with extra params.
- defaultCorrName() - 类 中的静态方法org.apache.spark.mllib.stat.correlation.CorrelationNames
-
- DefaultCredentials - org.apache.spark.streaming.kinesis中的类
-
Returns DefaultAWSCredentialsProviderChain for authentication.
- DefaultCredentials() - 类 的构造器org.apache.spark.streaming.kinesis.DefaultCredentials
-
- defaultLink() - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.Binomial$
-
- defaultLink() - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.Gamma$
-
- defaultLink() - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.Gaussian$
-
- defaultLink() - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.Poisson$
-
- defaultMinPartitions() - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
Default min number of partitions for Hadoop RDDs when not given by user
- defaultMinPartitions() - 类 中的方法org.apache.spark.SparkContext
-
Default min number of partitions for Hadoop RDDs when not given by user
Notice that we use math.min so the "defaultMinPartitions" cannot be higher than 2.
- defaultParallelism() - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
Default level of parallelism to use when not given by user (e.g. parallelize and makeRDD).
- defaultParallelism() - 类 中的方法org.apache.spark.SparkContext
-
Default level of parallelism to use when not given by user (e.g. parallelize and makeRDD).
- defaultParamMap() - 接口 中的方法org.apache.spark.ml.param.Params
-
Internal param map for default values.
- defaultParams(String) - 类 中的静态方法org.apache.spark.mllib.tree.configuration.BoostingStrategy
-
Returns default configuration for the boosting algorithm
- defaultParams(Enumeration.Value) - 类 中的静态方法org.apache.spark.mllib.tree.configuration.BoostingStrategy
-
Returns default configuration for the boosting algorithm
- DefaultParamsReadable<T> - org.apache.spark.ml.util中的接口
-
:: DeveloperApi ::
Helper trait for making simple Params types readable.
- DefaultParamsWritable - org.apache.spark.ml.util中的接口
-
:: DeveloperApi ::
Helper trait for making simple Params types writable.
- DefaultPartitionCoalescer - org.apache.spark.rdd中的类
-
Coalesce the partitions of a parent RDD (prev) into fewer partitions, so that each partition of
this RDD computes one or more of the parent ones.
- DefaultPartitionCoalescer(double) - 类 的构造器org.apache.spark.rdd.DefaultPartitionCoalescer
-
- DefaultPartitionCoalescer.PartitionLocations - org.apache.spark.rdd中的类
-
- defaultPartitioner(RDD<?>, Seq<RDD<?>>) - 类 中的静态方法org.apache.spark.Partitioner
-
Choose a partitioner to use for a cogroup-like operation between a number of RDDs.
- defaultSize() - 类 中的方法org.apache.spark.sql.types.ArrayType
-
The default size of a value of the ArrayType is the default size of the element type.
- defaultSize() - 类 中的方法org.apache.spark.sql.types.BinaryType
-
The default size of a value of the BinaryType is 100 bytes.
- defaultSize() - 类 中的方法org.apache.spark.sql.types.BooleanType
-
The default size of a value of the BooleanType is 1 byte.
- defaultSize() - 类 中的方法org.apache.spark.sql.types.ByteType
-
The default size of a value of the ByteType is 1 byte.
- defaultSize() - 类 中的方法org.apache.spark.sql.types.CalendarIntervalType
-
- defaultSize() - 类 中的静态方法org.apache.spark.sql.types.CharType
-
- defaultSize() - 类 中的方法org.apache.spark.sql.types.DataType
-
The default size of a value of this data type, used internally for size estimation.
- defaultSize() - 类 中的方法org.apache.spark.sql.types.DateType
-
The default size of a value of the DateType is 4 bytes.
- defaultSize() - 类 中的方法org.apache.spark.sql.types.DecimalType
-
The default size of a value of the DecimalType is 8 bytes when precision is at most 18,
and 16 bytes otherwise.
- defaultSize() - 类 中的方法org.apache.spark.sql.types.DoubleType
-
The default size of a value of the DoubleType is 8 bytes.
- defaultSize() - 类 中的方法org.apache.spark.sql.types.FloatType
-
The default size of a value of the FloatType is 4 bytes.
- defaultSize() - 类 中的方法org.apache.spark.sql.types.HiveStringType
-
- defaultSize() - 类 中的方法org.apache.spark.sql.types.IntegerType
-
The default size of a value of the IntegerType is 4 bytes.
- defaultSize() - 类 中的方法org.apache.spark.sql.types.LongType
-
The default size of a value of the LongType is 8 bytes.
- defaultSize() - 类 中的方法org.apache.spark.sql.types.MapType
-
The default size of a value of the MapType is
(the default size of the key type + the default size of the value type).
- defaultSize() - 类 中的方法org.apache.spark.sql.types.NullType
-
- defaultSize() - 类 中的静态方法org.apache.spark.sql.types.NumericType
-
- defaultSize() - 类 中的方法org.apache.spark.sql.types.ObjectType
-
- defaultSize() - 类 中的方法org.apache.spark.sql.types.ShortType
-
The default size of a value of the ShortType is 2 bytes.
- defaultSize() - 类 中的方法org.apache.spark.sql.types.StringType
-
The default size of a value of the StringType is 20 bytes.
- defaultSize() - 类 中的方法org.apache.spark.sql.types.StructType
-
The default size of a value of the StructType is the total default sizes of all field types.
- defaultSize() - 类 中的方法org.apache.spark.sql.types.TimestampType
-
The default size of a value of the TimestampType is 8 bytes.
- defaultSize() - 类 中的静态方法org.apache.spark.sql.types.VarcharType
-
- defaultStrategy(String) - 类 中的静态方法org.apache.spark.mllib.tree.configuration.Strategy
-
- defaultStrategy(Enumeration.Value) - 类 中的静态方法org.apache.spark.mllib.tree.configuration.Strategy
-
- DefaultTopologyMapper - org.apache.spark.storage中的类
-
A TopologyMapper that assumes all nodes are in the same rack
- DefaultTopologyMapper(SparkConf) - 类 的构造器org.apache.spark.storage.DefaultTopologyMapper
-
- defaultValue() - 类 中的方法org.apache.spark.internal.config.ConfigEntryWithDefault
-
- defaultValue() - 类 中的方法org.apache.spark.internal.config.ConfigEntryWithDefaultFunction
-
- defaultValue() - 类 中的方法org.apache.spark.internal.config.ConfigEntryWithDefaultString
-
- defaultValueString() - 类 中的方法org.apache.spark.internal.config.ConfigEntryWithDefault
-
- defaultValueString() - 类 中的方法org.apache.spark.internal.config.ConfigEntryWithDefaultFunction
-
- defaultValueString() - 类 中的方法org.apache.spark.internal.config.ConfigEntryWithDefaultString
-
- degree() - 类 中的方法org.apache.spark.ml.feature.PolynomialExpansion
-
The polynomial degree to expand, which should be greater than equal to 1.
- degrees() - 类 中的方法org.apache.spark.graphx.GraphOps
-
The degree of each vertex in the graph.
- degrees(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Converts an angle measured in radians to an approximately equivalent angle measured in degrees.
- degrees(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Converts an angle measured in radians to an approximately equivalent angle measured in degrees.
- degreesOfFreedom() - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegressionSummary
-
Degrees of freedom.
- degreesOfFreedom() - 类 中的方法org.apache.spark.ml.regression.LinearRegressionSummary
-
Degrees of freedom
- degreesOfFreedom() - 类 中的方法org.apache.spark.mllib.stat.test.ChiSqTestResult
-
- degreesOfFreedom() - 类 中的方法org.apache.spark.mllib.stat.test.KolmogorovSmirnovTestResult
-
- degreesOfFreedom() - 接口 中的方法org.apache.spark.mllib.stat.test.TestResult
-
Returns the degree(s) of freedom of the hypothesis test.
- delegate() - 类 中的方法org.apache.spark.InterruptibleIterator
-
- delete(Class<?>, Object) - 类 中的方法org.apache.spark.util.kvstore.InMemoryStore
-
- delete(Class<?>, Object) - 接口 中的方法org.apache.spark.util.kvstore.KVStore
-
Removes an object and all data related to it, like index entries, from the store.
- delete(Class<?>, Object) - 类 中的方法org.apache.spark.util.kvstore.LevelDB
-
- deleteCheckpointFiles() - 类 中的方法org.apache.spark.ml.clustering.DistributedLDAModel
-
:: DeveloperApi ::
Remove any remaining checkpoint files from training.
- deleteRecursively(File) - 类 中的静态方法org.apache.spark.util.Utils
-
Delete a file or directory and its contents recursively.
- deleteWithJob(FileSystem, Path, boolean) - 类 中的方法org.apache.spark.internal.io.FileCommitProtocol
-
Specifies that a file should be deleted with the commit of this job.
- deleteWithJob(FileSystem, Path, boolean) - 类 中的静态方法org.apache.spark.internal.io.HadoopMapReduceCommitProtocol
-
- delimiterOptions() - 类 中的静态方法org.apache.spark.sql.hive.execution.HiveOptions
-
- delta() - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.Tweedie$
-
Constant used in initialization and deviance to avoid numerical issues.
- dense(int, int, double[]) - 类 中的静态方法org.apache.spark.ml.linalg.Matrices
-
Creates a column-major dense matrix.
- dense(double, double...) - 类 中的静态方法org.apache.spark.ml.linalg.Vectors
-
Creates a dense vector from its values.
- dense(double, Seq<Object>) - 类 中的静态方法org.apache.spark.ml.linalg.Vectors
-
Creates a dense vector from its values.
- dense(double[]) - 类 中的静态方法org.apache.spark.ml.linalg.Vectors
-
Creates a dense vector from a double array.
- dense(int, int, double[]) - 类 中的静态方法org.apache.spark.mllib.linalg.Matrices
-
Creates a column-major dense matrix.
- dense(double, double...) - 类 中的静态方法org.apache.spark.mllib.linalg.Vectors
-
Creates a dense vector from its values.
- dense(double, Seq<Object>) - 类 中的静态方法org.apache.spark.mllib.linalg.Vectors
-
Creates a dense vector from its values.
- dense(double[]) - 类 中的静态方法org.apache.spark.mllib.linalg.Vectors
-
Creates a dense vector from a double array.
- dense_rank() - 类 中的静态方法org.apache.spark.sql.functions
-
Window function: returns the rank of rows within a window partition, without any gaps.
- DenseMatrix - org.apache.spark.ml.linalg中的类
-
Column-major dense matrix.
- DenseMatrix(int, int, double[], boolean) - 类 的构造器org.apache.spark.ml.linalg.DenseMatrix
-
- DenseMatrix(int, int, double[]) - 类 的构造器org.apache.spark.ml.linalg.DenseMatrix
-
Column-major dense matrix.
- DenseMatrix - org.apache.spark.mllib.linalg中的类
-
Column-major dense matrix.
- DenseMatrix(int, int, double[], boolean) - 类 的构造器org.apache.spark.mllib.linalg.DenseMatrix
-
- DenseMatrix(int, int, double[]) - 类 的构造器org.apache.spark.mllib.linalg.DenseMatrix
-
Column-major dense matrix.
- DenseVector - org.apache.spark.ml.linalg中的类
-
A dense vector represented by a value array.
- DenseVector(double[]) - 类 的构造器org.apache.spark.ml.linalg.DenseVector
-
- DenseVector - org.apache.spark.mllib.linalg中的类
-
A dense vector represented by a value array.
- DenseVector(double[]) - 类 的构造器org.apache.spark.mllib.linalg.DenseVector
-
- dependencies() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- dependencies() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- dependencies() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- dependencies() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- dependencies() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- dependencies() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- dependencies() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- dependencies() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- dependencies() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- dependencies() - 类 中的方法org.apache.spark.rdd.RDD
-
Get the list of dependencies of this RDD, taking into account whether the
RDD is checkpointed or not.
- dependencies() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- dependencies() - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
List of parent DStreams on which this DStream depends on
- dependencies() - 类 中的方法org.apache.spark.streaming.dstream.InputDStream
-
- Dependency<T> - org.apache.spark中的类
-
:: DeveloperApi ::
Base class for dependencies.
- Dependency() - 类 的构造器org.apache.spark.Dependency
-
- DEPLOY_MODE - 类 中的静态变量org.apache.spark.launcher.SparkLauncher
-
The Spark deploy mode.
- deployMode() - 类 中的方法org.apache.spark.SparkContext
-
- depth() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- depth() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- depth() - 类 中的方法org.apache.spark.mllib.tree.model.DecisionTreeModel
-
Get depth of tree.
- depth() - 类 中的方法org.apache.spark.util.sketch.CountMinSketch
-
- DerbyDialect - org.apache.spark.sql.jdbc中的类
-
- DerbyDialect() - 类 的构造器org.apache.spark.sql.jdbc.DerbyDialect
-
- deriv(double) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.CLogLog$
-
- deriv(double) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.Identity$
-
- deriv(double) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.Inverse$
-
- deriv(double) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.Log$
-
- deriv(double) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.Logit$
-
- deriv(double) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.Probit$
-
- deriv(double) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.Sqrt$
-
- desc() - 类 中的方法org.apache.spark.serializer.SerializationDebugger.ObjectStreamClassMethods
-
- desc() - 类 中的方法org.apache.spark.sql.Column
-
Returns a sort expression based on the descending order of the column.
- desc(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns a sort expression based on the descending order of the column.
- desc() - 类 中的方法org.apache.spark.util.MethodIdentifier
-
- desc_nulls_first() - 类 中的方法org.apache.spark.sql.Column
-
Returns a sort expression based on the descending order of the column,
and null values appear before non-null values.
- desc_nulls_first(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns a sort expression based on the descending order of the column,
and null values appear before non-null values.
- desc_nulls_last() - 类 中的方法org.apache.spark.sql.Column
-
Returns a sort expression based on the descending order of the column,
and null values appear after non-null values.
- desc_nulls_last(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns a sort expression based on the descending order of the column,
and null values appear after non-null values.
- describe(String...) - 类 中的方法org.apache.spark.sql.Dataset
-
Computes basic statistics for numeric and string columns, including count, mean, stddev, min,
and max.
- describe(Seq<String>) - 类 中的方法org.apache.spark.sql.Dataset
-
Computes basic statistics for numeric and string columns, including count, mean, stddev, min,
and max.
- describeTopics(int) - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- describeTopics() - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- describeTopics(int) - 类 中的方法org.apache.spark.ml.clustering.LDAModel
-
Return the topics described by their top-weighted terms.
- describeTopics() - 类 中的方法org.apache.spark.ml.clustering.LDAModel
-
- describeTopics(int) - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- describeTopics() - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- describeTopics(int) - 类 中的方法org.apache.spark.mllib.clustering.DistributedLDAModel
-
- describeTopics(int) - 类 中的方法org.apache.spark.mllib.clustering.LDAModel
-
Return the topics described by weighted terms.
- describeTopics() - 类 中的方法org.apache.spark.mllib.clustering.LDAModel
-
Return the topics described by weighted terms.
- describeTopics(int) - 类 中的方法org.apache.spark.mllib.clustering.LocalLDAModel
-
- description() - 类 中的方法org.apache.spark.ExceptionFailure
-
- description() - 类 中的方法org.apache.spark.sql.catalog.Column
-
- description() - 类 中的方法org.apache.spark.sql.catalog.Database
-
- description() - 类 中的方法org.apache.spark.sql.catalog.Function
-
- description() - 类 中的方法org.apache.spark.sql.catalog.Table
-
- description() - 类 中的方法org.apache.spark.sql.streaming.SinkProgress
-
- description() - 类 中的方法org.apache.spark.sql.streaming.SourceProgress
-
- description() - 类 中的方法org.apache.spark.status.api.v1.JobData
-
- description() - 类 中的方法org.apache.spark.status.api.v1.StageData
-
- description() - 类 中的方法org.apache.spark.status.api.v1.streaming.OutputOperationInfo
-
- description() - 类 中的方法org.apache.spark.status.LiveStage
-
- description() - 类 中的方法org.apache.spark.storage.StorageLevel
-
- description() - 类 中的方法org.apache.spark.streaming.scheduler.OutputOperationInfo
-
- DESER_CPU_TIME() - 类 中的静态方法org.apache.spark.status.TaskIndexNames
-
- DESER_TIME() - 类 中的静态方法org.apache.spark.status.TaskIndexNames
-
- DeserializationStream - org.apache.spark.serializer中的类
-
:: DeveloperApi ::
A stream for reading serialized objects.
- DeserializationStream() - 类 的构造器org.apache.spark.serializer.DeserializationStream
-
- deserialize(Object) - 类 中的方法org.apache.spark.mllib.linalg.VectorUDT
-
- deserialize(ByteBuffer, ClassLoader, ClassTag<T>) - 类 中的方法org.apache.spark.serializer.DummySerializerInstance
-
- deserialize(ByteBuffer, ClassTag<T>) - 类 中的方法org.apache.spark.serializer.DummySerializerInstance
-
- deserialize(ByteBuffer, ClassTag<T>) - 类 中的方法org.apache.spark.serializer.SerializerInstance
-
- deserialize(ByteBuffer, ClassLoader, ClassTag<T>) - 类 中的方法org.apache.spark.serializer.SerializerInstance
-
- deserialize(byte[], Class<T>) - 类 中的方法org.apache.spark.util.kvstore.KVStoreSerializer
-
- deserialize(byte[]) - 类 中的静态方法org.apache.spark.util.Utils
-
Deserialize an object using Java serialization
- deserialize(byte[], ClassLoader) - 类 中的静态方法org.apache.spark.util.Utils
-
Deserialize an object using Java serialization and the given ClassLoader
- deserialized() - 类 中的方法org.apache.spark.storage.StorageLevel
-
- DeserializedMemoryEntry<T> - org.apache.spark.storage.memory中的类
-
- DeserializedMemoryEntry(Object, long, ClassTag<T>) - 类 的构造器org.apache.spark.storage.memory.DeserializedMemoryEntry
-
- deserializeLongValue(byte[]) - 类 中的静态方法org.apache.spark.util.Utils
-
Deserialize a Long value (used for PythonPartitioner)
- deserializeOffset(String) - 接口 中的方法org.apache.spark.sql.sources.v2.reader.streaming.ContinuousReader
-
Deserialize a JSON string into an Offset of the implementation-defined offset type.
- deserializeOffset(String) - 接口 中的方法org.apache.spark.sql.sources.v2.reader.streaming.MicroBatchReader
-
Deserialize a JSON string into an Offset of the implementation-defined offset type.
- deserializeStream(InputStream) - 类 中的方法org.apache.spark.serializer.DummySerializerInstance
-
- deserializeStream(InputStream) - 类 中的方法org.apache.spark.serializer.SerializerInstance
-
- deserializeViaNestedStream(InputStream, SerializerInstance, Function1<DeserializationStream, BoxedUnit>) - 类 中的静态方法org.apache.spark.util.Utils
-
Deserialize via nested stream using specific serializer
- destroy() - 类 中的方法org.apache.spark.broadcast.Broadcast
-
Destroy all data and metadata related to this broadcast variable.
- details() - 类 中的方法org.apache.spark.scheduler.StageInfo
-
- details() - 类 中的方法org.apache.spark.status.api.v1.StageData
-
- determineBounds(ArrayBuffer<Tuple2<K, Object>>, int, Ordering<K>, ClassTag<K>) - 类 中的静态方法org.apache.spark.RangePartitioner
-
Determines the bounds for range partitioning from candidates with weights indicating how many
items each represents.
- DetermineTableStats - org.apache.spark.sql.hive中的类
-
- DetermineTableStats(SparkSession) - 类 的构造器org.apache.spark.sql.hive.DetermineTableStats
-
- deterministic() - 类 中的方法org.apache.spark.sql.expressions.UserDefinedAggregateFunction
-
Returns true iff this function is deterministic, i.e. given the same input,
always return the same output.
- deterministic() - 类 中的方法org.apache.spark.sql.expressions.UserDefinedFunction
-
Returns true iff the UDF is deterministic, i.e. the UDF produces the same output given the same
input.
- deviance(double, double, double) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.Binomial$
-
- deviance(double, double, double) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.Gamma$
-
- deviance(double, double, double) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.Gaussian$
-
- deviance(double, double, double) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.Poisson$
-
- deviance() - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegressionSummary
-
The deviance for the fitted model.
- devianceResiduals() - 类 中的方法org.apache.spark.ml.regression.LinearRegressionSummary
-
The weighted residuals, the usual residuals rescaled by
the square root of the instance weights.
- dfToCols(Dataset<Row>) - 类 中的静态方法org.apache.spark.sql.api.r.SQLUtils
-
- dfToRowRDD(Dataset<Row>) - 类 中的静态方法org.apache.spark.sql.api.r.SQLUtils
-
- dgemm(double, DenseMatrix<Object>, DenseMatrix<Object>, double, DenseMatrix<Object>) - 类 中的静态方法org.apache.spark.ml.ann.BreezeUtil
-
DGEMM: C := alpha * A * B + beta * C
- dgemv(double, DenseMatrix<Object>, DenseVector<Object>, double, DenseVector<Object>) - 类 中的静态方法org.apache.spark.ml.ann.BreezeUtil
-
DGEMV: y := alpha * A * x + beta * y
- diag(Vector) - 类 中的静态方法org.apache.spark.ml.linalg.DenseMatrix
-
Generate a diagonal matrix in DenseMatrix format from the supplied values.
- diag(Vector) - 类 中的静态方法org.apache.spark.ml.linalg.Matrices
-
Generate a diagonal matrix in Matrix format from the supplied values.
- diag(Vector) - 类 中的静态方法org.apache.spark.mllib.linalg.DenseMatrix
-
Generate a diagonal matrix in DenseMatrix format from the supplied values.
- diag(Vector) - 类 中的静态方法org.apache.spark.mllib.linalg.Matrices
-
Generate a diagonal matrix in Matrix format from the supplied values.
- diff(RDD<Tuple2<Object, VD>>) - 类 中的方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- diff(VertexRDD<VD>) - 类 中的方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- diff(RDD<Tuple2<Object, VD>>) - 类 中的方法org.apache.spark.graphx.VertexRDD
-
For each vertex present in both this and other, diff returns only those vertices with
differing values; for values that are different, keeps the values from other.
- diff(VertexRDD<VD>) - 类 中的方法org.apache.spark.graphx.VertexRDD
-
For each vertex present in both this and other, diff returns only those vertices with
differing values; for values that are different, keeps the values from other.
- diff(GenSeq<B>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- dir() - 类 中的方法org.apache.spark.mllib.optimization.NNLS.Workspace
-
- directory(File) - 类 中的方法org.apache.spark.launcher.SparkLauncher
-
Sets the working directory of spark-submit.
- disableOutputSpecValidation() - 类 中的静态方法org.apache.spark.internal.io.SparkHadoopWriterUtils
-
Allows for the spark.hadoop.validateOutputSpecs checks to be disabled on a case-by-case
basis; see SPARK-4835 for more details.
- disconnect() - 接口 中的方法org.apache.spark.launcher.SparkAppHandle
-
Disconnects the handle from the application, without stopping it.
- DISK_BYTES_SPILLED() - 类 中的静态方法org.apache.spark.InternalAccumulator
-
- DISK_ONLY - 类 中的静态变量org.apache.spark.api.java.StorageLevels
-
- DISK_ONLY() - 类 中的静态方法org.apache.spark.storage.StorageLevel
-
- DISK_ONLY_2 - 类 中的静态变量org.apache.spark.api.java.StorageLevels
-
- DISK_ONLY_2() - 类 中的静态方法org.apache.spark.storage.StorageLevel
-
- DISK_SPILL() - 类 中的静态方法org.apache.spark.status.TaskIndexNames
-
- DiskBlockData - org.apache.spark.storage中的类
-
- DiskBlockData(long, long, File, long) - 类 的构造器org.apache.spark.storage.DiskBlockData
-
- diskBytesSpilled() - 类 中的方法org.apache.spark.status.api.v1.ExecutorStageSummary
-
- diskBytesSpilled() - 类 中的方法org.apache.spark.status.api.v1.StageData
-
- diskBytesSpilled() - 类 中的方法org.apache.spark.status.api.v1.TaskMetricDistributions
-
- diskBytesSpilled() - 类 中的方法org.apache.spark.status.api.v1.TaskMetrics
-
- diskSize() - 类 中的方法org.apache.spark.storage.BlockManagerMessages.UpdateBlockInfo
-
- diskSize() - 类 中的方法org.apache.spark.storage.BlockStatus
-
- diskSize() - 类 中的方法org.apache.spark.storage.BlockUpdatedInfo
-
- diskSize() - 类 中的方法org.apache.spark.storage.RDDInfo
-
- diskUsed() - 类 中的方法org.apache.spark.status.api.v1.ExecutorSummary
-
- diskUsed() - 类 中的方法org.apache.spark.status.api.v1.RDDDataDistribution
-
- diskUsed() - 类 中的方法org.apache.spark.status.api.v1.RDDPartitionInfo
-
- diskUsed() - 类 中的方法org.apache.spark.status.api.v1.RDDStorageInfo
-
- diskUsed() - 类 中的方法org.apache.spark.status.LiveExecutor
-
- diskUsed() - 类 中的方法org.apache.spark.status.LiveRDD
-
- diskUsed() - 类 中的方法org.apache.spark.status.LiveRDDDistribution
-
- diskUsed() - 类 中的方法org.apache.spark.status.LiveRDDPartition
-
- diskUsed() - 类 中的方法org.apache.spark.storage.StorageStatus
-
已过时。
Return the disk space used by this block manager.
- diskUsedByRdd(int) - 类 中的方法org.apache.spark.storage.StorageStatus
-
已过时。
Return the disk space used by the given RDD in this block manager in O(1) time.
- dispersion() - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegressionSummary
-
The dispersion of the fitted model.
- dispose() - 类 中的方法org.apache.spark.storage.DiskBlockData
-
- dispose() - 类 中的方法org.apache.spark.storage.EncryptedBlockData
-
- dispose(ByteBuffer) - 类 中的静态方法org.apache.spark.storage.StorageUtils
-
Attempt to clean up a ByteBuffer if it is direct or memory-mapped.
- distinct() - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
Return a new RDD containing the distinct elements in this RDD.
- distinct(int) - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
Return a new RDD containing the distinct elements in this RDD.
- distinct() - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Return a new RDD containing the distinct elements in this RDD.
- distinct(int) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Return a new RDD containing the distinct elements in this RDD.
- distinct() - 类 中的方法org.apache.spark.api.java.JavaRDD
-
Return a new RDD containing the distinct elements in this RDD.
- distinct(int) - 类 中的方法org.apache.spark.api.java.JavaRDD
-
Return a new RDD containing the distinct elements in this RDD.
- distinct(int, Ordering<T>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- distinct() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- distinct(int, Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- distinct() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- distinct(int, Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- distinct() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- distinct(int, Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- distinct() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- distinct(int, Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- distinct() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- distinct(int, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- distinct() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- distinct(int, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- distinct() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- distinct(int, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- distinct() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- distinct(int, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- distinct() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- distinct(int, Ordering<T>) - 类 中的方法org.apache.spark.rdd.RDD
-
Return a new RDD containing the distinct elements in this RDD.
- distinct() - 类 中的方法org.apache.spark.rdd.RDD
-
Return a new RDD containing the distinct elements in this RDD.
- distinct(int, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- distinct() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- distinct() - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new Dataset that contains only the unique rows from this Dataset.
- distinct(Column...) - 类 中的方法org.apache.spark.sql.expressions.UserDefinedAggregateFunction
-
Creates a Column for this UDAF using the distinct values of the given
Columns as input arguments.
- distinct(Seq<Column>) - 类 中的方法org.apache.spark.sql.expressions.UserDefinedAggregateFunction
-
Creates a Column for this UDAF using the distinct values of the given
Columns as input arguments.
- distinct() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- distinct$default$2(int) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- distinct$default$2(int) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- distinct$default$2(int) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- distinct$default$2(int) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- distinct$default$2(int) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- distinct$default$2(int) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- distinct$default$2(int) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- distinct$default$2(int) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- distinct$default$2(int) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- distinct$default$2(int) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- DistributedLDAModel - org.apache.spark.ml.clustering中的类
-
Distributed model fitted by
LDA.
- DistributedLDAModel - org.apache.spark.mllib.clustering中的类
-
Distributed LDA model.
- DistributedMatrix - org.apache.spark.mllib.linalg.distributed中的接口
-
Represents a distributively stored matrix backed by one or more RDDs.
- Distribution - org.apache.spark.sql.sources.v2.reader.partitioning中的接口
-
An interface to represent data distribution requirement, which specifies how the records should
be distributed among the data partitions(one
DataReader outputs data for one partition).
- distribution(LiveExecutor) - 类 中的方法org.apache.spark.status.LiveRDD
-
- distributionOpt(LiveExecutor) - 类 中的方法org.apache.spark.status.LiveRDD
-
- div(Decimal, Decimal) - 类 中的方法org.apache.spark.sql.types.Decimal.DecimalIsFractional$
-
- div(Duration) - 类 中的方法org.apache.spark.streaming.Duration
-
- divide(Object) - 类 中的方法org.apache.spark.sql.Column
-
Division this expression by another expression.
- doc() - 类 中的静态方法org.apache.spark.ml.param.DoubleParam
-
- doc() - 类 中的静态方法org.apache.spark.ml.param.FloatParam
-
- doc() - 类 中的方法org.apache.spark.ml.param.Param
-
- docConcentration() - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- docConcentration() - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- docConcentration() - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- docConcentration() - 类 中的方法org.apache.spark.mllib.clustering.DistributedLDAModel
-
- docConcentration() - 类 中的方法org.apache.spark.mllib.clustering.LDAModel
-
Concentration parameter (commonly named "alpha") for the prior placed on documents'
distributions over topics ("theta").
- docConcentration() - 类 中的方法org.apache.spark.mllib.clustering.LocalLDAModel
-
- DocumentFrequencyAggregator(int) - 类 的构造器org.apache.spark.mllib.feature.IDF.DocumentFrequencyAggregator
-
- DocumentFrequencyAggregator() - 类 的构造器org.apache.spark.mllib.feature.IDF.DocumentFrequencyAggregator
-
- doesDirectoryContainAnyNewFiles(File, long) - 类 中的静态方法org.apache.spark.util.Utils
-
Determines if a directory contains any files newer than cutoff seconds.
- doFetchFile(String, File, String, SparkConf, org.apache.spark.SecurityManager, Configuration) - 类 中的静态方法org.apache.spark.util.Utils
-
Download a file or directory to target directory.
- Dot - org.apache.spark.ml.feature中的类
-
- Dot() - 类 的构造器org.apache.spark.ml.feature.Dot
-
- dot(Vector, Vector) - 类 中的静态方法org.apache.spark.ml.linalg.BLAS
-
dot(x, y)
- dot(Vector, Vector) - 类 中的静态方法org.apache.spark.mllib.linalg.BLAS
-
dot(x, y)
- doTest(DStream<Tuple2<StatCounter, StatCounter>>) - 类 中的静态方法org.apache.spark.mllib.stat.test.StudentTTest
-
- doTest(DStream<Tuple2<StatCounter, StatCounter>>) - 类 中的静态方法org.apache.spark.mllib.stat.test.WelchTTest
-
- DOUBLE() - 类 中的静态方法org.apache.spark.sql.Encoders
-
An encoder for nullable double type.
- doubleAccumulator(double) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
- doubleAccumulator(double, String) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
- doubleAccumulator() - 类 中的方法org.apache.spark.SparkContext
-
Create and register a double accumulator, which starts with 0 and accumulates inputs by add.
- doubleAccumulator(String) - 类 中的方法org.apache.spark.SparkContext
-
Create and register a double accumulator, which starts with 0 and accumulates inputs by add.
- DoubleAccumulator - org.apache.spark.util中的类
-
An
accumulator for computing sum, count, and averages for double precision
floating numbers.
- DoubleAccumulator() - 类 的构造器org.apache.spark.util.DoubleAccumulator
-
- DoubleAccumulatorParam$() - 类 的构造器org.apache.spark.AccumulatorParam.DoubleAccumulatorParam$
-
已过时。
- DoubleArrayArrayParam - org.apache.spark.ml.param中的类
-
:: DeveloperApi ::
Specialized version of Param[Array[Array[Double}] for Java.
- DoubleArrayArrayParam(Params, String, String, Function1<double[][], Object>) - 类 的构造器org.apache.spark.ml.param.DoubleArrayArrayParam
-
- DoubleArrayArrayParam(Params, String, String) - 类 的构造器org.apache.spark.ml.param.DoubleArrayArrayParam
-
- DoubleArrayParam - org.apache.spark.ml.param中的类
-
:: DeveloperApi ::
Specialized version of Param[Array[Double} for Java.
- DoubleArrayParam(Params, String, String, Function1<double[], Object>) - 类 的构造器org.apache.spark.ml.param.DoubleArrayParam
-
- DoubleArrayParam(Params, String, String) - 类 的构造器org.apache.spark.ml.param.DoubleArrayParam
-
- DoubleFlatMapFunction<T> - org.apache.spark.api.java.function中的接口
-
A function that returns zero or more records of type Double from each input record.
- DoubleFunction<T> - org.apache.spark.api.java.function中的接口
-
A function that returns Doubles, and can be used to construct DoubleRDDs.
- DoubleParam - org.apache.spark.ml.param中的类
-
:: DeveloperApi ::
Specialized version of Param[Double] for Java.
- DoubleParam(String, String, String, Function1<Object, Object>) - 类 的构造器org.apache.spark.ml.param.DoubleParam
-
- DoubleParam(String, String, String) - 类 的构造器org.apache.spark.ml.param.DoubleParam
-
- DoubleParam(Identifiable, String, String, Function1<Object, Object>) - 类 的构造器org.apache.spark.ml.param.DoubleParam
-
- DoubleParam(Identifiable, String, String) - 类 的构造器org.apache.spark.ml.param.DoubleParam
-
- DoubleRDDFunctions - org.apache.spark.rdd中的类
-
Extra functions available on RDDs of Doubles through an implicit conversion.
- DoubleRDDFunctions(RDD<Object>) - 类 的构造器org.apache.spark.rdd.DoubleRDDFunctions
-
- doubleRDDToDoubleRDDFunctions(RDD<Object>) - 类 中的静态方法org.apache.spark.rdd.RDD
-
- DoubleType - 类 中的静态变量org.apache.spark.sql.types.DataTypes
-
Gets the DoubleType object.
- DoubleType - org.apache.spark.sql.types中的类
-
The data type representing Double values.
- driver() - 类 中的方法org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.SetupDriver
-
- DRIVER_EXTRA_CLASSPATH - 类 中的静态变量org.apache.spark.launcher.SparkLauncher
-
Configuration key for the driver class path.
- DRIVER_EXTRA_JAVA_OPTIONS - 类 中的静态变量org.apache.spark.launcher.SparkLauncher
-
Configuration key for the driver VM options.
- DRIVER_EXTRA_LIBRARY_PATH - 类 中的静态变量org.apache.spark.launcher.SparkLauncher
-
Configuration key for the driver native library path.
- DRIVER_MEMORY - 类 中的静态变量org.apache.spark.launcher.SparkLauncher
-
Configuration key for the driver memory.
- DRIVER_WAL_BATCHING_CONF_KEY() - 类 中的静态方法org.apache.spark.streaming.util.WriteAheadLogUtils
-
- DRIVER_WAL_BATCHING_TIMEOUT_CONF_KEY() - 类 中的静态方法org.apache.spark.streaming.util.WriteAheadLogUtils
-
- DRIVER_WAL_CLASS_CONF_KEY() - 类 中的静态方法org.apache.spark.streaming.util.WriteAheadLogUtils
-
- DRIVER_WAL_CLOSE_AFTER_WRITE_CONF_KEY() - 类 中的静态方法org.apache.spark.streaming.util.WriteAheadLogUtils
-
- DRIVER_WAL_MAX_FAILURES_CONF_KEY() - 类 中的静态方法org.apache.spark.streaming.util.WriteAheadLogUtils
-
- DRIVER_WAL_ROLLING_INTERVAL_CONF_KEY() - 类 中的静态方法org.apache.spark.streaming.util.WriteAheadLogUtils
-
- driverLogs() - 类 中的方法org.apache.spark.scheduler.SparkListenerApplicationStart
-
- drop() - 类 中的方法org.apache.spark.sql.DataFrameNaFunctions
-
Returns a new DataFrame that drops rows containing any null or NaN values.
- drop(String) - 类 中的方法org.apache.spark.sql.DataFrameNaFunctions
-
Returns a new DataFrame that drops rows containing null or NaN values.
- drop(String[]) - 类 中的方法org.apache.spark.sql.DataFrameNaFunctions
-
Returns a new DataFrame that drops rows containing any null or NaN values
in the specified columns.
- drop(Seq<String>) - 类 中的方法org.apache.spark.sql.DataFrameNaFunctions
-
(Scala-specific) Returns a new DataFrame that drops rows containing any null or NaN values
in the specified columns.
- drop(String, String[]) - 类 中的方法org.apache.spark.sql.DataFrameNaFunctions
-
Returns a new DataFrame that drops rows containing null or NaN values
in the specified columns.
- drop(String, Seq<String>) - 类 中的方法org.apache.spark.sql.DataFrameNaFunctions
-
(Scala-specific) Returns a new DataFrame that drops rows containing null or NaN values
in the specified columns.
- drop(int) - 类 中的方法org.apache.spark.sql.DataFrameNaFunctions
-
Returns a new DataFrame that drops rows containing
less than minNonNulls non-null and non-NaN values.
- drop(int, String[]) - 类 中的方法org.apache.spark.sql.DataFrameNaFunctions
-
Returns a new DataFrame that drops rows containing
less than minNonNulls non-null and non-NaN values in the specified columns.
- drop(int, Seq<String>) - 类 中的方法org.apache.spark.sql.DataFrameNaFunctions
-
(Scala-specific) Returns a new DataFrame that drops rows containing less than
minNonNulls non-null and non-NaN values in the specified columns.
- drop(String...) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new Dataset with columns dropped.
- drop(String) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new Dataset with a column dropped.
- drop(Seq<String>) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new Dataset with columns dropped.
- drop(Column) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new Dataset with a column dropped.
- drop(int) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- dropDuplicates(String, String...) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new
Dataset with duplicate rows removed, considering only
the subset of columns.
- dropDuplicates() - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new Dataset that contains only the unique rows from this Dataset.
- dropDuplicates(Seq<String>) - 类 中的方法org.apache.spark.sql.Dataset
-
(Scala-specific) Returns a new Dataset with duplicate rows removed, considering only
the subset of columns.
- dropDuplicates(String[]) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new Dataset with duplicate rows removed, considering only
the subset of columns.
- dropDuplicates(String, Seq<String>) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new
Dataset with duplicate rows removed, considering only
the subset of columns.
- dropGlobalTempView(String) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
Drops the global temporary view with the given view name in the catalog.
- dropLast() - 类 中的方法org.apache.spark.ml.feature.OneHotEncoder
-
已过时。
Whether to drop the last category in the encoded vector (default: true)
- dropLast() - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderEstimator
-
- dropLast() - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderModel
-
- dropRight(int) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- dropTempTable(String) - 类 中的方法org.apache.spark.sql.SQLContext
-
- dropTempView(String) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
Drops the local temporary view with the given view name in the catalog.
- dropWhile(Function1<A, Object>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- dspmv(int, double, DenseVector, DenseVector, double, DenseVector) - 类 中的静态方法org.apache.spark.ml.linalg.BLAS
-
y := alpha*A*x + beta*y
- Dst - 类 中的静态变量org.apache.spark.graphx.TripletFields
-
Expose the destination and edge fields but not the source field.
- dstAttr() - 类 中的方法org.apache.spark.graphx.EdgeContext
-
The vertex attribute of the edge's destination vertex.
- dstAttr() - 类 中的方法org.apache.spark.graphx.EdgeTriplet
-
The destination vertex attribute
- dstAttr() - 类 中的方法org.apache.spark.graphx.impl.AggregatingEdgeContext
-
- dstId() - 类 中的方法org.apache.spark.graphx.Edge
-
- dstId() - 类 中的方法org.apache.spark.graphx.EdgeContext
-
The vertex id of the edge's destination vertex.
- dstId() - 类 中的方法org.apache.spark.graphx.impl.AggregatingEdgeContext
-
- dstream() - 类 中的方法org.apache.spark.streaming.api.java.JavaDStream
-
- dstream() - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
- dstream() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- dstream() - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- dstream() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- dstream() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- dstream() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- DStream<T> - org.apache.spark.streaming.dstream中的类
-
A Discretized Stream (DStream), the basic abstraction in Spark Streaming, is a continuous
sequence of RDDs (of the same type) representing a continuous stream of data (see
org.apache.spark.rdd.RDD in the Spark core documentation for more details on RDDs).
- DStream(StreamingContext, ClassTag<T>) - 类 的构造器org.apache.spark.streaming.dstream.DStream
-
- dtypes() - 类 中的方法org.apache.spark.sql.Dataset
-
Returns all column names and their data types as an array.
- DummySerializerInstance - org.apache.spark.serializer中的类
-
Unfortunately, we need a serializer instance in order to construct a DiskBlockObjectWriter.
- duration() - 类 中的方法org.apache.spark.scheduler.TaskInfo
-
- duration() - 类 中的方法org.apache.spark.status.api.v1.ApplicationAttemptInfo
-
- duration() - 类 中的方法org.apache.spark.status.api.v1.streaming.OutputOperationInfo
-
- duration() - 类 中的方法org.apache.spark.status.api.v1.TaskData
-
- DURATION() - 类 中的静态方法org.apache.spark.status.TaskIndexNames
-
- Duration - org.apache.spark.streaming中的类
-
- Duration(long) - 类 的构造器org.apache.spark.streaming.Duration
-
- duration() - 类 中的方法org.apache.spark.streaming.scheduler.OutputOperationInfo
-
Return the duration of this output operation.
- durationMs() - 类 中的方法org.apache.spark.sql.streaming.StreamingQueryProgress
-
- Durations - org.apache.spark.streaming中的类
-
- Durations() - 类 的构造器org.apache.spark.streaming.Durations
-
- f() - 类 中的方法org.apache.spark.sql.expressions.UserDefinedFunction
-
- f1Measure() - 类 中的方法org.apache.spark.mllib.evaluation.MultilabelMetrics
-
Returns document-based f1-measure averaged by the number of documents
- f1Measure(double) - 类 中的方法org.apache.spark.mllib.evaluation.MultilabelMetrics
-
Returns f1-measure for a given label (category)
- factorial(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Computes the factorial of the given value.
- failed() - 类 中的方法org.apache.spark.scheduler.TaskInfo
-
- FAILED() - 类 中的静态方法org.apache.spark.TaskState
-
- failedStages() - 类 中的方法org.apache.spark.status.LiveJob
-
- failedTasks() - 类 中的方法org.apache.spark.status.api.v1.ExecutorStageSummary
-
- failedTasks() - 类 中的方法org.apache.spark.status.api.v1.ExecutorSummary
-
- failedTasks() - 类 中的方法org.apache.spark.status.LiveExecutor
-
- failedTasks() - 类 中的方法org.apache.spark.status.LiveExecutorStageSummary
-
- failedTasks() - 类 中的方法org.apache.spark.status.LiveJob
-
- failedTasks() - 类 中的方法org.apache.spark.status.LiveStage
-
- failure(String) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaParser
-
- failureReason() - 类 中的方法org.apache.spark.scheduler.StageInfo
-
If the stage failed, the reason why.
- failureReason() - 类 中的方法org.apache.spark.status.api.v1.StageData
-
- failureReason() - 类 中的方法org.apache.spark.status.api.v1.streaming.OutputOperationInfo
-
- failureReason() - 类 中的方法org.apache.spark.streaming.scheduler.OutputOperationInfo
-
- failureReasonCell(String, int, boolean) - 类 中的静态方法org.apache.spark.streaming.ui.UIUtils
-
- FAIR() - 类 中的静态方法org.apache.spark.scheduler.SchedulingMode
-
- FAKE_HIVE_VERSION() - 类 中的静态方法org.apache.spark.sql.hive.HiveUtils
-
- FalsePositiveRate - org.apache.spark.mllib.evaluation.binary中的类
-
False positive rate.
- FalsePositiveRate() - 类 的构造器org.apache.spark.mllib.evaluation.binary.FalsePositiveRate
-
- falsePositiveRate(double) - 类 中的方法org.apache.spark.mllib.evaluation.MulticlassMetrics
-
Returns false positive rate for a given label (category)
- falsePositiveRateByLabel() - 接口 中的方法org.apache.spark.ml.classification.LogisticRegressionSummary
-
Returns false positive rate for each label (category).
- family() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- family() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- family() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- family() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- Family$() - 类 的构造器org.apache.spark.ml.regression.GeneralizedLinearRegression.Family$
-
- FamilyAndLink$() - 类 的构造器org.apache.spark.ml.regression.GeneralizedLinearRegression.FamilyAndLink$
-
- fastEquals(TreeNode<?>) - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- fastEquals(TreeNode<?>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- fastEquals(TreeNode<?>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- fastEquals(TreeNode<?>) - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- fdr() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelector
-
- fdr() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- fdr() - 类 中的方法org.apache.spark.mllib.feature.ChiSqSelector
-
- feature() - 类 中的方法org.apache.spark.mllib.feature.ChiSqSelectorModel.SaveLoadV1_0$.Data
-
- feature() - 类 中的方法org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.SplitData
-
- feature() - 类 中的方法org.apache.spark.mllib.tree.model.Split
-
- FeatureHasher - org.apache.spark.ml.feature中的类
-
Feature hashing projects a set of categorical or numerical features into a feature vector of
specified dimension (typically substantially smaller than that of the original feature
space).
- FeatureHasher(String) - 类 的构造器org.apache.spark.ml.feature.FeatureHasher
-
- FeatureHasher() - 类 的构造器org.apache.spark.ml.feature.FeatureHasher
-
- featureImportances() - 类 中的方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
Estimate of the importance of each feature.
- featureImportances() - 类 中的方法org.apache.spark.ml.classification.GBTClassificationModel
-
Estimate of the importance of each feature.
- featureImportances() - 类 中的方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
Estimate of the importance of each feature.
- featureImportances() - 类 中的方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
Estimate of the importance of each feature.
- featureImportances() - 类 中的方法org.apache.spark.ml.regression.GBTRegressionModel
-
Estimate of the importance of each feature.
- featureImportances() - 类 中的方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
Estimate of the importance of each feature.
- featureIndex() - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegression
-
- featureIndex() - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegressionModel
-
- featureIndex() - 类 中的方法org.apache.spark.ml.tree.CategoricalSplit
-
- featureIndex() - 类 中的方法org.apache.spark.ml.tree.ContinuousSplit
-
- featureIndex() - 类 中的方法org.apache.spark.ml.tree.DecisionTreeModelReadWrite.SplitData
-
- featureIndex() - 接口 中的方法org.apache.spark.ml.tree.Split
-
Index of feature which this split tests
- features() - 类 中的方法org.apache.spark.ml.feature.LabeledPoint
-
- features() - 类 中的方法org.apache.spark.mllib.regression.LabeledPoint
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- featuresCol() - 接口 中的方法org.apache.spark.ml.classification.LogisticRegressionSummary
-
Field in "predictions" which gives the features of each instance as a vector.
- featuresCol() - 类 中的方法org.apache.spark.ml.classification.LogisticRegressionSummaryImpl
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.classification.OneVsRest
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.classification.OneVsRestModel
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeans
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
- featuresCol() - 类 中的方法org.apache.spark.ml.clustering.ClusteringSummary
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixture
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.clustering.KMeans
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.clustering.KMeansModel
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.evaluation.ClusteringEvaluator
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelector
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.feature.RFormula
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.feature.RFormulaModel
-
- featuresCol() - 接口 中的方法org.apache.spark.ml.param.shared.HasFeaturesCol
-
Param for features column name.
- featuresCol() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegression
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegressionModel
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- featuresCol() - 类 中的方法org.apache.spark.ml.regression.LinearRegressionSummary
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- featuresCol() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- featureSubsetStrategy() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- featureSubsetStrategy() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- featureSubsetStrategy() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- featureSubsetStrategy() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- featureSubsetStrategy() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- featureSubsetStrategy() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- featureSubsetStrategy() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- featureSubsetStrategy() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- featureSum() - 类 中的方法org.apache.spark.ml.evaluation.SquaredEuclideanSilhouette.ClusterStats
-
- FeatureType - org.apache.spark.mllib.tree.configuration中的类
-
Enum to describe whether a feature is "continuous" or "categorical"
- FeatureType() - 类 的构造器org.apache.spark.mllib.tree.configuration.FeatureType
-
- featureType() - 类 中的方法org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.SplitData
-
- featureType() - 类 中的方法org.apache.spark.mllib.tree.model.Split
-
- FETCH_WAIT_TIME() - 类 中的方法org.apache.spark.InternalAccumulator.shuffleRead$
-
- FetchFailed - org.apache.spark中的类
-
:: DeveloperApi ::
Task failed to fetch shuffle data from a remote node.
- FetchFailed(BlockManagerId, int, int, int, String) - 类 的构造器org.apache.spark.FetchFailed
-
- fetchFile(String, File, SparkConf, org.apache.spark.SecurityManager, Configuration, long, boolean) - 类 中的静态方法org.apache.spark.util.Utils
-
Download a file or directory to target directory.
- fetchPct() - 类 中的方法org.apache.spark.scheduler.RuntimePercentage
-
- fetchWaitTime() - 类 中的方法org.apache.spark.status.api.v1.ShuffleReadMetricDistributions
-
- fetchWaitTime() - 类 中的方法org.apache.spark.status.api.v1.ShuffleReadMetrics
-
- field() - 类 中的方法org.apache.spark.storage.BroadcastBlockId
-
- fieldIndex(String) - 接口 中的方法org.apache.spark.sql.Row
-
Returns the index of a given field name.
- fieldIndex(String) - 类 中的方法org.apache.spark.sql.types.StructType
-
Returns the index of a given field.
- fieldNames() - 类 中的方法org.apache.spark.sql.types.StructType
-
Returns all field names in an array.
- fields() - 类 中的方法org.apache.spark.sql.types.StructType
-
- FIFO() - 类 中的静态方法org.apache.spark.scheduler.SchedulingMode
-
- FILE_FORMAT() - 类 中的静态方法org.apache.spark.sql.hive.execution.HiveOptions
-
- FileBasedTopologyMapper - org.apache.spark.storage中的类
-
A simple file based topology mapper.
- FileBasedTopologyMapper(SparkConf) - 类 的构造器org.apache.spark.storage.FileBasedTopologyMapper
-
- FileCommitProtocol - org.apache.spark.internal.io中的类
-
An interface to define how a single Spark job commits its outputs.
- FileCommitProtocol() - 类 的构造器org.apache.spark.internal.io.FileCommitProtocol
-
- FileCommitProtocol.EmptyTaskCommitMessage$ - org.apache.spark.internal.io中的类
-
- FileCommitProtocol.TaskCommitMessage - org.apache.spark.internal.io中的类
-
- fileFormat() - 类 中的方法org.apache.spark.sql.hive.execution.HiveOptions
-
- files() - 类 中的方法org.apache.spark.SparkContext
-
- fileStream(String, Class<K>, Class<V>, Class<F>) - 类 中的方法org.apache.spark.streaming.api.java.JavaStreamingContext
-
Create an input stream that monitors a Hadoop-compatible filesystem
for new files and reads them using the given key-value types and input format.
- fileStream(String, Class<K>, Class<V>, Class<F>, Function<Path, Boolean>, boolean) - 类 中的方法org.apache.spark.streaming.api.java.JavaStreamingContext
-
Create an input stream that monitors a Hadoop-compatible filesystem
for new files and reads them using the given key-value types and input format.
- fileStream(String, Class<K>, Class<V>, Class<F>, Function<Path, Boolean>, boolean, Configuration) - 类 中的方法org.apache.spark.streaming.api.java.JavaStreamingContext
-
Create an input stream that monitors a Hadoop-compatible filesystem
for new files and reads them using the given key-value types and input format.
- fileStream(String, ClassTag<K>, ClassTag<V>, ClassTag<F>) - 类 中的方法org.apache.spark.streaming.StreamingContext
-
Create an input stream that monitors a Hadoop-compatible filesystem
for new files and reads them using the given key-value types and input format.
- fileStream(String, Function1<Path, Object>, boolean, ClassTag<K>, ClassTag<V>, ClassTag<F>) - 类 中的方法org.apache.spark.streaming.StreamingContext
-
Create an input stream that monitors a Hadoop-compatible filesystem
for new files and reads them using the given key-value types and input format.
- fileStream(String, Function1<Path, Object>, boolean, Configuration, ClassTag<K>, ClassTag<V>, ClassTag<F>) - 类 中的方法org.apache.spark.streaming.StreamingContext
-
Create an input stream that monitors a Hadoop-compatible filesystem
for new files and reads them using the given key-value types and input format.
- fill(long) - 类 中的方法org.apache.spark.sql.DataFrameNaFunctions
-
Returns a new DataFrame that replaces null or NaN values in numeric columns with value.
- fill(double) - 类 中的方法org.apache.spark.sql.DataFrameNaFunctions
-
Returns a new DataFrame that replaces null or NaN values in numeric columns with value.
- fill(String) - 类 中的方法org.apache.spark.sql.DataFrameNaFunctions
-
Returns a new DataFrame that replaces null values in string columns with value.
- fill(long, String[]) - 类 中的方法org.apache.spark.sql.DataFrameNaFunctions
-
Returns a new DataFrame that replaces null or NaN values in specified numeric columns.
- fill(double, String[]) - 类 中的方法org.apache.spark.sql.DataFrameNaFunctions
-
Returns a new DataFrame that replaces null or NaN values in specified numeric columns.
- fill(long, Seq<String>) - 类 中的方法org.apache.spark.sql.DataFrameNaFunctions
-
(Scala-specific) Returns a new DataFrame that replaces null or NaN values in specified
numeric columns.
- fill(double, Seq<String>) - 类 中的方法org.apache.spark.sql.DataFrameNaFunctions
-
(Scala-specific) Returns a new DataFrame that replaces null or NaN values in specified
numeric columns.
- fill(String, String[]) - 类 中的方法org.apache.spark.sql.DataFrameNaFunctions
-
Returns a new DataFrame that replaces null values in specified string columns.
- fill(String, Seq<String>) - 类 中的方法org.apache.spark.sql.DataFrameNaFunctions
-
(Scala-specific) Returns a new DataFrame that replaces null values in
specified string columns.
- fill(boolean) - 类 中的方法org.apache.spark.sql.DataFrameNaFunctions
-
Returns a new DataFrame that replaces null values in boolean columns with value.
- fill(boolean, Seq<String>) - 类 中的方法org.apache.spark.sql.DataFrameNaFunctions
-
(Scala-specific) Returns a new DataFrame that replaces null values in specified
boolean columns.
- fill(boolean, String[]) - 类 中的方法org.apache.spark.sql.DataFrameNaFunctions
-
Returns a new DataFrame that replaces null values in specified boolean columns.
- fill(Map<String, Object>) - 类 中的方法org.apache.spark.sql.DataFrameNaFunctions
-
Returns a new DataFrame that replaces null values.
- fill(Map<String, Object>) - 类 中的方法org.apache.spark.sql.DataFrameNaFunctions
-
(Scala-specific) Returns a new DataFrame that replaces null values.
- fillInStackTrace() - 异常错误 中的静态方法org.apache.spark.sql.AnalysisException
-
- filter(Function<Double, Boolean>) - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
Return a new RDD containing only the elements that satisfy a predicate.
- filter(Function<Tuple2<K, V>, Boolean>) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Return a new RDD containing only the elements that satisfy a predicate.
- filter(Function<T, Boolean>) - 类 中的方法org.apache.spark.api.java.JavaRDD
-
Return a new RDD containing only the elements that satisfy a predicate.
- filter(Function1<T, Object>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- filter(Function1<T, Object>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- filter(Function1<Graph<VD, ED>, Graph<VD2, ED2>>, Function1<EdgeTriplet<VD2, ED2>, Object>, Function2<Object, VD2, Object>, ClassTag<VD2>, ClassTag<ED2>) - 类 中的方法org.apache.spark.graphx.GraphOps
-
Filter the graph by computing some values to filter on, and applying the predicates.
- filter(Function1<EdgeTriplet<VD, ED>, Object>, Function2<Object, VD, Object>) - 类 中的方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- filter(Function1<Tuple2<Object, VD>, Object>) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- filter(Function1<Tuple2<Object, VD>, Object>) - 类 中的方法org.apache.spark.graphx.VertexRDD
-
Restricts the vertex set to the set of vertices satisfying the given predicate.
- filter(Params) - 类 中的方法org.apache.spark.ml.param.ParamMap
-
Filters this param map for the given parent.
- filter(Function1<T, Object>) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- filter(Function1<T, Object>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- filter(Function1<T, Object>) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- filter(Function1<T, Object>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- filter(Function1<T, Object>) - 类 中的方法org.apache.spark.rdd.RDD
-
Return a new RDD containing only the elements that satisfy a predicate.
- filter(Function1<T, Object>) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- filter(Column) - 类 中的方法org.apache.spark.sql.Dataset
-
Filters rows using the given condition.
- filter(String) - 类 中的方法org.apache.spark.sql.Dataset
-
Filters rows using the given SQL expression.
- filter(Function1<T, Object>) - 类 中的方法org.apache.spark.sql.Dataset
-
:: Experimental ::
(Scala-specific)
Returns a new Dataset that only contains elements where func returns true.
- filter(FilterFunction<T>) - 类 中的方法org.apache.spark.sql.Dataset
-
:: Experimental ::
(Java-specific)
Returns a new Dataset that only contains elements where func returns true.
- Filter - org.apache.spark.sql.sources中的类
-
A filter predicate for data sources.
- Filter() - 类 的构造器org.apache.spark.sql.sources.Filter
-
- filter(Function1<A, Object>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- filter() - 类 中的方法org.apache.spark.storage.BlockManagerMessages.GetMatchingBlockIds
-
- filter(Function<T, Boolean>) - 类 中的方法org.apache.spark.streaming.api.java.JavaDStream
-
Return a new DStream containing only the elements that satisfy a predicate.
- filter(Function<T, Boolean>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- filter(Function<Tuple2<K, V>, Boolean>) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream containing only the elements that satisfy a predicate.
- filter(Function<Tuple2<K, V>, Boolean>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- filter(Function<Tuple2<K, V>, Boolean>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- filter(Function<T, Boolean>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- filter(Function1<T, Object>) - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Return a new DStream containing only the elements that satisfy a predicate.
- filterByRange(K, K) - 类 中的方法org.apache.spark.rdd.OrderedRDDFunctions
-
Returns an RDD containing only the elements in the inclusive range lower to upper.
- FilterFunction<T> - org.apache.spark.api.java.function中的接口
-
Base interface for a function used in Dataset's filter function.
- filterName() - 类 中的方法org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.AddWebUIFilter
-
- filterNot(Function1<A, Object>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- filterParams() - 类 中的方法org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.AddWebUIFilter
-
- finalStorageLevel() - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- find(Function1<BaseType, Object>) - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- find(Function1<BaseType, Object>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- find(Function1<BaseType, Object>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- find(Function1<BaseType, Object>) - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- find(Function1<A, Object>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- findMissingPartitions() - 类 中的方法org.apache.spark.ShuffleStatus
-
Returns the sequence of partition ids that are missing (i.e. needs to be computed).
- findSynonyms(String, int) - 类 中的方法org.apache.spark.ml.feature.Word2VecModel
-
Find "num" number of words closest in similarity to the given word, not
including the word itself.
- findSynonyms(Vector, int) - 类 中的方法org.apache.spark.ml.feature.Word2VecModel
-
Find "num" number of words whose vector representation is most similar to the supplied vector.
- findSynonyms(String, int) - 类 中的方法org.apache.spark.mllib.feature.Word2VecModel
-
Find synonyms of a word; do not include the word itself in results.
- findSynonyms(Vector, int) - 类 中的方法org.apache.spark.mllib.feature.Word2VecModel
-
Find synonyms of the vector representation of a word, possibly
including any words in the model vocabulary whose vector respresentation
is the supplied vector.
- findSynonymsArray(Vector, int) - 类 中的方法org.apache.spark.ml.feature.Word2VecModel
-
Find "num" number of words whose vector representation is most similar to the supplied vector.
- findSynonymsArray(String, int) - 类 中的方法org.apache.spark.ml.feature.Word2VecModel
-
Find "num" number of words closest in similarity to the given word, not
including the word itself.
- finish(BUF) - 类 中的方法org.apache.spark.sql.expressions.Aggregator
-
Transform the output of the reduction.
- finished() - 类 中的方法org.apache.spark.scheduler.TaskInfo
-
- FINISHED() - 类 中的静态方法org.apache.spark.TaskState
-
- finishTime() - 类 中的方法org.apache.spark.scheduler.TaskInfo
-
The time when the task has completed successfully (including the time to remotely fetch
results, if necessary).
- first() - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
- first() - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
- first() - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- first() - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Return the first element in this RDD.
- first() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- first() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- first() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- first() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- first() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- first() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- first() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- first() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- first() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- first() - 类 中的方法org.apache.spark.rdd.RDD
-
Return the first element in this RDD.
- first() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- first() - 类 中的方法org.apache.spark.sql.Dataset
-
Returns the first row.
- first(Column, boolean) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the first value in a group.
- first(String, boolean) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the first value of a column in a group.
- first(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the first value in a group.
- first(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the first value of a column in a group.
- first(Object) - 类 中的方法org.apache.spark.util.kvstore.KVStoreView
-
Iterates starting at the given value of the chosen index (inclusive).
- firstFailureReason() - 类 中的方法org.apache.spark.status.api.v1.streaming.BatchInfo
-
- firstLaunchTime() - 类 中的方法org.apache.spark.status.LiveStage
-
- firstTaskLaunchedTime() - 类 中的方法org.apache.spark.status.api.v1.StageData
-
- fit(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- fit(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- fit(Dataset<?>, ParamMap[]) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- fit(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- fit(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- fit(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- fit(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- fit(Dataset<?>, ParamMap[]) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- fit(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- fit(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- fit(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- fit(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- fit(Dataset<?>, ParamMap[]) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- fit(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- fit(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- fit(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- fit(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- fit(Dataset<?>, ParamMap[]) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- fit(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- fit(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- fit(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- fit(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- fit(Dataset<?>, ParamMap[]) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- fit(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- fit(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- fit(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- fit(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- fit(Dataset<?>, ParamMap[]) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- fit(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- fit(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- fit(Dataset<?>) - 类 中的方法org.apache.spark.ml.classification.OneVsRest
-
- fit(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- fit(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- fit(Dataset<?>, ParamMap[]) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- fit(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- fit(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- fit(Dataset<?>) - 类 中的方法org.apache.spark.ml.clustering.BisectingKMeans
-
- fit(Dataset<?>) - 类 中的方法org.apache.spark.ml.clustering.GaussianMixture
-
- fit(Dataset<?>) - 类 中的方法org.apache.spark.ml.clustering.KMeans
-
- fit(Dataset<?>) - 类 中的方法org.apache.spark.ml.clustering.LDA
-
- fit(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的方法org.apache.spark.ml.Estimator
-
Fits a single model to the input data with optional parameters.
- fit(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的方法org.apache.spark.ml.Estimator
-
Fits a single model to the input data with optional parameters.
- fit(Dataset<?>, ParamMap) - 类 中的方法org.apache.spark.ml.Estimator
-
Fits a single model to the input data with provided parameter map.
- fit(Dataset<?>) - 类 中的方法org.apache.spark.ml.Estimator
-
Fits a model to the input data.
- fit(Dataset<?>, ParamMap[]) - 类 中的方法org.apache.spark.ml.Estimator
-
Fits multiple models to the input data with multiple sets of parameters.
- fit(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- fit(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- fit(Dataset<?>, ParamMap[]) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- fit(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- fit(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- fit(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.ChiSqSelector
-
- fit(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.CountVectorizer
-
- fit(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.IDF
-
- fit(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.Imputer
-
- fit(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.MaxAbsScaler
-
- fit(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSH
-
- fit(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSH
-
- fit(Dataset<?>, ParamMap[]) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSH
-
- fit(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSH
-
- fit(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSH
-
- fit(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.MinMaxScaler
-
- fit(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.OneHotEncoderEstimator
-
- fit(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.PCA
-
Computes a
PCAModel that contains the principal components of the input vectors.
- fit(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.QuantileDiscretizer
-
- fit(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.RFormula
-
- fit(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.StandardScaler
-
- fit(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.StringIndexer
-
- fit(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.VectorIndexer
-
- fit(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.Word2Vec
-
- fit(Dataset<?>) - 类 中的方法org.apache.spark.ml.fpm.FPGrowth
-
- fit(Dataset<?>) - 类 中的方法org.apache.spark.ml.Pipeline
-
Fits the pipeline to the input dataset with additional parameters.
- fit(Dataset<?>) - 类 中的方法org.apache.spark.ml.Predictor
-
- fit(Dataset<?>) - 类 中的方法org.apache.spark.ml.recommendation.ALS
-
- fit(Dataset<?>) - 类 中的方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- fit(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- fit(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- fit(Dataset<?>, ParamMap[]) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- fit(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- fit(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- fit(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- fit(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- fit(Dataset<?>, ParamMap[]) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- fit(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- fit(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- fit(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- fit(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- fit(Dataset<?>, ParamMap[]) - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- fit(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- fit(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- fit(Dataset<?>) - 类 中的方法org.apache.spark.ml.regression.IsotonicRegression
-
- fit(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- fit(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- fit(Dataset<?>, ParamMap[]) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- fit(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- fit(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- fit(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- fit(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- fit(Dataset<?>, ParamMap[]) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- fit(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- fit(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- fit(Dataset<?>) - 类 中的方法org.apache.spark.ml.tuning.CrossValidator
-
- fit(Dataset<?>) - 类 中的方法org.apache.spark.ml.tuning.TrainValidationSplit
-
- fit(RDD<LabeledPoint>) - 类 中的方法org.apache.spark.mllib.feature.ChiSqSelector
-
Returns a ChiSquared feature selector.
- fit(RDD<Vector>) - 类 中的方法org.apache.spark.mllib.feature.IDF
-
Computes the inverse document frequency.
- fit(JavaRDD<Vector>) - 类 中的方法org.apache.spark.mllib.feature.IDF
-
Computes the inverse document frequency.
- fit(RDD<Vector>) - 类 中的方法org.apache.spark.mllib.feature.PCA
-
Computes a
PCAModel that contains the principal components of the input vectors.
- fit(JavaRDD<Vector>) - 类 中的方法org.apache.spark.mllib.feature.PCA
-
Java-friendly version of fit().
- fit(RDD<Vector>) - 类 中的方法org.apache.spark.mllib.feature.StandardScaler
-
Computes the mean and variance and stores as a model to be used for later scaling.
- fit(RDD<S>) - 类 中的方法org.apache.spark.mllib.feature.Word2Vec
-
Computes the vector representation of each word in vocabulary.
- fit(JavaRDD<S>) - 类 中的方法org.apache.spark.mllib.feature.Word2Vec
-
Computes the vector representation of each word in vocabulary (Java version).
- fitIntercept() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- fitIntercept() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- fitIntercept() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- fitIntercept() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- fitIntercept() - 接口 中的方法org.apache.spark.ml.param.shared.HasFitIntercept
-
Param for whether to fit an intercept term.
- fitIntercept() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- fitIntercept() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- fitIntercept() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- fitIntercept() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- fitIntercept() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- fitIntercept() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- Fixed$() - 类 的构造器org.apache.spark.sql.types.DecimalType.Fixed$
-
- flatMap(FlatMapFunction<T, U>) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- flatMap(FlatMapFunction<T, U>) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- flatMap(FlatMapFunction<T, U>) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- flatMap(FlatMapFunction<T, U>) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Return a new RDD by first applying a function to all elements of this
RDD, and then flattening the results.
- flatMap(Function1<T, TraversableOnce<U>>, ClassTag<U>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- flatMap(Function1<T, TraversableOnce<U>>, ClassTag<U>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- flatMap(Function1<T, TraversableOnce<U>>, ClassTag<U>) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- flatMap(Function1<T, TraversableOnce<U>>, ClassTag<U>) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- flatMap(Function1<T, TraversableOnce<U>>, ClassTag<U>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- flatMap(Function1<T, TraversableOnce<U>>, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- flatMap(Function1<T, TraversableOnce<U>>, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- flatMap(Function1<T, TraversableOnce<U>>, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- flatMap(Function1<T, TraversableOnce<U>>, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- flatMap(Function1<T, TraversableOnce<U>>, ClassTag<U>) - 类 中的方法org.apache.spark.rdd.RDD
-
Return a new RDD by first applying a function to all elements of this
RDD, and then flattening the results.
- flatMap(Function1<T, TraversableOnce<U>>, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- flatMap(Function1<T, TraversableOnce<U>>, Encoder<U>) - 类 中的方法org.apache.spark.sql.Dataset
-
:: Experimental ::
(Scala-specific)
Returns a new Dataset by first applying a function to all elements of this Dataset,
and then flattening the results.
- flatMap(FlatMapFunction<T, U>, Encoder<U>) - 类 中的方法org.apache.spark.sql.Dataset
-
:: Experimental ::
(Java-specific)
Returns a new Dataset by first applying a function to all elements of this Dataset,
and then flattening the results.
- flatMap(Function1<BaseType, TraversableOnce<A>>) - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- flatMap(Function1<BaseType, TraversableOnce<A>>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- flatMap(Function1<BaseType, TraversableOnce<A>>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- flatMap(Function1<BaseType, TraversableOnce<A>>) - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- flatMap(Function1<A, GenTraversableOnce<B>>, CanBuildFrom<Repr, B, That>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- flatMap(FlatMapFunction<T, U>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- flatMap(FlatMapFunction<T, U>) - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return a new DStream by applying a function to all elements of this DStream,
and then flattening the results
- flatMap(FlatMapFunction<T, U>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- flatMap(FlatMapFunction<T, U>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- flatMap(FlatMapFunction<T, U>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- flatMap(FlatMapFunction<T, U>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- flatMap(FlatMapFunction<T, U>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- flatMap(Function1<T, TraversableOnce<U>>, ClassTag<U>) - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Return a new DStream by applying a function to all elements of this DStream,
and then flattening the results
- FlatMapFunction<T,R> - org.apache.spark.api.java.function中的接口
-
A function that returns zero or more output records from each input record.
- FlatMapFunction2<T1,T2,R> - org.apache.spark.api.java.function中的接口
-
A function that takes two inputs and returns zero or more output records.
- flatMapGroups(Function2<K, Iterator<V>, TraversableOnce<U>>, Encoder<U>) - 类 中的方法org.apache.spark.sql.KeyValueGroupedDataset
-
(Scala-specific)
Applies the given function to each group of data.
- flatMapGroups(FlatMapGroupsFunction<K, V, U>, Encoder<U>) - 类 中的方法org.apache.spark.sql.KeyValueGroupedDataset
-
(Java-specific)
Applies the given function to each group of data.
- FlatMapGroupsFunction<K,V,R> - org.apache.spark.api.java.function中的接口
-
A function that returns zero or more output records from each grouping key and its values.
- flatMapGroupsWithState(OutputMode, GroupStateTimeout, Function3<K, Iterator<V>, GroupState<S>, Iterator<U>>, Encoder<S>, Encoder<U>) - 类 中的方法org.apache.spark.sql.KeyValueGroupedDataset
-
::Experimental::
(Scala-specific)
Applies the given function to each group of data, while maintaining a user-defined per-group
state.
- flatMapGroupsWithState(FlatMapGroupsWithStateFunction<K, V, S, U>, OutputMode, Encoder<S>, Encoder<U>, GroupStateTimeout) - 类 中的方法org.apache.spark.sql.KeyValueGroupedDataset
-
::Experimental::
(Java-specific)
Applies the given function to each group of data, while maintaining a user-defined per-group
state.
- FlatMapGroupsWithStateFunction<K,V,S,R> - org.apache.spark.api.java.function中的接口
-
::Experimental::
Base interface for a map function used in
org.apache.spark.sql.KeyValueGroupedDataset.flatMapGroupsWithState(
FlatMapGroupsWithStateFunction, org.apache.spark.sql.streaming.OutputMode,
org.apache.spark.sql.Encoder, org.apache.spark.sql.Encoder)
- flatMapToDouble(DoubleFlatMapFunction<T>) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- flatMapToDouble(DoubleFlatMapFunction<T>) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- flatMapToDouble(DoubleFlatMapFunction<T>) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- flatMapToDouble(DoubleFlatMapFunction<T>) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Return a new RDD by first applying a function to all elements of this
RDD, and then flattening the results.
- flatMapToPair(PairFlatMapFunction<T, K2, V2>) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- flatMapToPair(PairFlatMapFunction<T, K2, V2>) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- flatMapToPair(PairFlatMapFunction<T, K2, V2>) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- flatMapToPair(PairFlatMapFunction<T, K2, V2>) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Return a new RDD by first applying a function to all elements of this
RDD, and then flattening the results.
- flatMapToPair(PairFlatMapFunction<T, K2, V2>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- flatMapToPair(PairFlatMapFunction<T, K2, V2>) - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return a new DStream by applying a function to all elements of this DStream,
and then flattening the results
- flatMapToPair(PairFlatMapFunction<T, K2, V2>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- flatMapToPair(PairFlatMapFunction<T, K2, V2>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- flatMapToPair(PairFlatMapFunction<T, K2, V2>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- flatMapToPair(PairFlatMapFunction<T, K2, V2>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- flatMapToPair(PairFlatMapFunction<T, K2, V2>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- flatMapValues(Function<V, Iterable<U>>) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Pass each value in the key-value pair RDD through a flatMap function without changing the
keys; this also retains the original RDD's partitioning.
- flatMapValues(Function1<V, TraversableOnce<U>>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Pass each value in the key-value pair RDD through a flatMap function without changing the
keys; this also retains the original RDD's partitioning.
- flatMapValues(Function<V, Iterable<U>>) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying a flatmap function to the value of each key-value pairs in
'this' DStream without changing the key.
- flatMapValues(Function<V, Iterable<U>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- flatMapValues(Function<V, Iterable<U>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- flatMapValues(Function1<V, TraversableOnce<U>>, ClassTag<U>) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying a flatmap function to the value of each key-value pairs in
'this' DStream without changing the key.
- flatten(Function1<A, GenTraversableOnce<B>>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- FLOAT() - 类 中的静态方法org.apache.spark.sql.Encoders
-
An encoder for nullable float type.
- FloatAccumulatorParam$() - 类 的构造器org.apache.spark.AccumulatorParam.FloatAccumulatorParam$
-
已过时。
- FloatParam - org.apache.spark.ml.param中的类
-
:: DeveloperApi ::
Specialized version of Param[Float] for Java.
- FloatParam(String, String, String, Function1<Object, Object>) - 类 的构造器org.apache.spark.ml.param.FloatParam
-
- FloatParam(String, String, String) - 类 的构造器org.apache.spark.ml.param.FloatParam
-
- FloatParam(Identifiable, String, String, Function1<Object, Object>) - 类 的构造器org.apache.spark.ml.param.FloatParam
-
- FloatParam(Identifiable, String, String) - 类 的构造器org.apache.spark.ml.param.FloatParam
-
- FloatType - 类 中的静态变量org.apache.spark.sql.types.DataTypes
-
Gets the FloatType object.
- FloatType - org.apache.spark.sql.types中的类
-
The data type representing Float values.
- floor(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Computes the floor of the given value.
- floor(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Computes the floor of the given column.
- floor() - 类 中的方法org.apache.spark.sql.types.Decimal
-
- floor(Duration) - 类 中的方法org.apache.spark.streaming.Time
-
- floor(Duration, Time) - 类 中的方法org.apache.spark.streaming.Time
-
- flush() - 类 中的方法org.apache.spark.io.SnappyOutputStreamWrapper
-
- flush() - 类 中的方法org.apache.spark.serializer.SerializationStream
-
- flush() - 类 中的方法org.apache.spark.storage.TimeTrackingOutputStream
-
- fMeasure(double, double) - 类 中的方法org.apache.spark.mllib.evaluation.MulticlassMetrics
-
Returns f-measure for a given label (category)
- fMeasure(double) - 类 中的方法org.apache.spark.mllib.evaluation.MulticlassMetrics
-
Returns f1-measure for a given label (category)
- fMeasure() - 类 中的方法org.apache.spark.mllib.evaluation.MulticlassMetrics
-
- fMeasureByLabel(double) - 接口 中的方法org.apache.spark.ml.classification.LogisticRegressionSummary
-
Returns f-measure for each label (category).
- fMeasureByLabel() - 接口 中的方法org.apache.spark.ml.classification.LogisticRegressionSummary
-
Returns f1-measure for each label (category).
- fMeasureByThreshold() - 接口 中的方法org.apache.spark.ml.classification.BinaryLogisticRegressionSummary
-
Returns a dataframe with two fields (threshold, F-Measure) curve with beta = 1.0.
- fMeasureByThreshold(double) - 类 中的方法org.apache.spark.mllib.evaluation.BinaryClassificationMetrics
-
Returns the (threshold, F-Measure) curve.
- fMeasureByThreshold() - 类 中的方法org.apache.spark.mllib.evaluation.BinaryClassificationMetrics
-
Returns the (threshold, F-Measure) curve with beta = 1.0.
- fold(T, Function2<T, T, T>) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- fold(T, Function2<T, T, T>) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- fold(T, Function2<T, T, T>) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- fold(T, Function2<T, T, T>) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Aggregate the elements of each partition, and then the results for all the partitions, using a
given associative function and a neutral "zero value".
- fold(T, Function2<T, T, T>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- fold(T, Function2<T, T, T>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- fold(T, Function2<T, T, T>) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- fold(T, Function2<T, T, T>) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- fold(T, Function2<T, T, T>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- fold(T, Function2<T, T, T>) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- fold(T, Function2<T, T, T>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- fold(T, Function2<T, T, T>) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- fold(T, Function2<T, T, T>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- fold(T, Function2<T, T, T>) - 类 中的方法org.apache.spark.rdd.RDD
-
Aggregate the elements of each partition, and then the results for all the partitions, using a
given associative function and a neutral "zero value".
- fold(T, Function2<T, T, T>) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- fold(A1, Function2<A1, A1, A1>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- foldByKey(V, Partitioner, Function2<V, V, V>) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Merge the values for each key using an associative function and a neutral "zero value" which
may be added to the result an arbitrary number of times, and must not change the result
(e.g ., Nil for list concatenation, 0 for addition, or 1 for multiplication.).
- foldByKey(V, int, Function2<V, V, V>) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Merge the values for each key using an associative function and a neutral "zero value" which
may be added to the result an arbitrary number of times, and must not change the result
(e.g ., Nil for list concatenation, 0 for addition, or 1 for multiplication.).
- foldByKey(V, Function2<V, V, V>) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Merge the values for each key using an associative function and a neutral "zero value"
which may be added to the result an arbitrary number of times, and must not change the result
(e.g., Nil for list concatenation, 0 for addition, or 1 for multiplication.).
- foldByKey(V, Partitioner, Function2<V, V, V>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Merge the values for each key using an associative function and a neutral "zero value" which
may be added to the result an arbitrary number of times, and must not change the result
(e.g., Nil for list concatenation, 0 for addition, or 1 for multiplication.).
- foldByKey(V, int, Function2<V, V, V>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Merge the values for each key using an associative function and a neutral "zero value" which
may be added to the result an arbitrary number of times, and must not change the result
(e.g., Nil for list concatenation, 0 for addition, or 1 for multiplication.).
- foldByKey(V, Function2<V, V, V>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Merge the values for each key using an associative function and a neutral "zero value" which
may be added to the result an arbitrary number of times, and must not change the result
(e.g., Nil for list concatenation, 0 for addition, or 1 for multiplication.).
- foldLeft(B, Function2<B, A, B>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- foldRight(B, Function2<A, B, B>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- forall(Function1<A, Object>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- forceIndexLabel() - 类 中的静态方法org.apache.spark.ml.feature.RFormula
-
- forceIndexLabel() - 类 中的静态方法org.apache.spark.ml.feature.RFormulaModel
-
- foreach(VoidFunction<T>) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- foreach(VoidFunction<T>) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- foreach(VoidFunction<T>) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- foreach(VoidFunction<T>) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Applies a function f to all elements of this RDD.
- foreach(Function1<T, BoxedUnit>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- foreach(Function1<T, BoxedUnit>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- foreach(Function1<T, BoxedUnit>) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- foreach(Function1<T, BoxedUnit>) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- foreach(Function1<T, BoxedUnit>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- foreach(Function1<T, BoxedUnit>) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- foreach(Function1<T, BoxedUnit>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- foreach(Function1<T, BoxedUnit>) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- foreach(Function1<T, BoxedUnit>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- foreach(Function1<T, BoxedUnit>) - 类 中的方法org.apache.spark.rdd.RDD
-
Applies a function f to all elements of this RDD.
- foreach(Function1<T, BoxedUnit>) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- foreach(Function1<T, BoxedUnit>) - 类 中的方法org.apache.spark.sql.Dataset
-
Applies a function f to all rows.
- foreach(ForeachFunction<T>) - 类 中的方法org.apache.spark.sql.Dataset
-
(Java-specific)
Runs func on each element of this Dataset.
- foreach(Function1<BaseType, BoxedUnit>) - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- foreach(Function1<BaseType, BoxedUnit>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- foreach(Function1<BaseType, BoxedUnit>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- foreach(Function1<BaseType, BoxedUnit>) - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- foreach(ForeachWriter<T>) - 类 中的方法org.apache.spark.sql.streaming.DataStreamWriter
-
Starts the execution of the streaming query, which will continually send results to the given
ForeachWriter as new data arrives.
- foreach(Function1<A, U>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- foreach(Function1<A, U>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- foreachActive(Function3<Object, Object, Object, BoxedUnit>) - 类 中的方法org.apache.spark.ml.linalg.DenseMatrix
-
- foreachActive(Function2<Object, Object, BoxedUnit>) - 类 中的方法org.apache.spark.ml.linalg.DenseVector
-
- foreachActive(Function3<Object, Object, Object, BoxedUnit>) - 接口 中的方法org.apache.spark.ml.linalg.Matrix
-
Applies a function f to all the active elements of dense and sparse matrix.
- foreachActive(Function3<Object, Object, Object, BoxedUnit>) - 类 中的方法org.apache.spark.ml.linalg.SparseMatrix
-
- foreachActive(Function2<Object, Object, BoxedUnit>) - 类 中的方法org.apache.spark.ml.linalg.SparseVector
-
- foreachActive(Function2<Object, Object, BoxedUnit>) - 接口 中的方法org.apache.spark.ml.linalg.Vector
-
Applies a function f to all the active elements of dense and sparse vector.
- foreachActive(Function2<Object, Object, BoxedUnit>) - 类 中的方法org.apache.spark.mllib.linalg.DenseVector
-
- foreachActive(Function3<Object, Object, Object, BoxedUnit>) - 接口 中的方法org.apache.spark.mllib.linalg.Matrix
-
Applies a function f to all the active elements of dense and sparse matrix.
- foreachActive(Function2<Object, Object, BoxedUnit>) - 类 中的方法org.apache.spark.mllib.linalg.SparseVector
-
- foreachActive(Function2<Object, Object, BoxedUnit>) - 接口 中的方法org.apache.spark.mllib.linalg.Vector
-
Applies a function f to all the active elements of dense and sparse vector.
- foreachAsync(VoidFunction<T>) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- foreachAsync(VoidFunction<T>) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- foreachAsync(VoidFunction<T>) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- foreachAsync(VoidFunction<T>) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
The asynchronous version of the foreach action, which
applies a function f to all the elements of this RDD.
- foreachAsync(Function1<T, BoxedUnit>) - 类 中的方法org.apache.spark.rdd.AsyncRDDActions
-
Applies a function f to all elements of this RDD.
- ForeachFunction<T> - org.apache.spark.api.java.function中的接口
-
Base interface for a function used in Dataset's foreach function.
- foreachPartition(VoidFunction<Iterator<T>>) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- foreachPartition(VoidFunction<Iterator<T>>) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- foreachPartition(VoidFunction<Iterator<T>>) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- foreachPartition(VoidFunction<Iterator<T>>) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Applies a function f to each partition of this RDD.
- foreachPartition(Function1<Iterator<T>, BoxedUnit>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- foreachPartition(Function1<Iterator<T>, BoxedUnit>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- foreachPartition(Function1<Iterator<T>, BoxedUnit>) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- foreachPartition(Function1<Iterator<T>, BoxedUnit>) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- foreachPartition(Function1<Iterator<T>, BoxedUnit>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- foreachPartition(Function1<Iterator<T>, BoxedUnit>) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- foreachPartition(Function1<Iterator<T>, BoxedUnit>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- foreachPartition(Function1<Iterator<T>, BoxedUnit>) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- foreachPartition(Function1<Iterator<T>, BoxedUnit>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- foreachPartition(Function1<Iterator<T>, BoxedUnit>) - 类 中的方法org.apache.spark.rdd.RDD
-
Applies a function f to each partition of this RDD.
- foreachPartition(Function1<Iterator<T>, BoxedUnit>) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- foreachPartition(Function1<Iterator<T>, BoxedUnit>) - 类 中的方法org.apache.spark.sql.Dataset
-
Applies a function f to each partition of this Dataset.
- foreachPartition(ForeachPartitionFunction<T>) - 类 中的方法org.apache.spark.sql.Dataset
-
(Java-specific)
Runs func on each partition of this Dataset.
- foreachPartitionAsync(VoidFunction<Iterator<T>>) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- foreachPartitionAsync(VoidFunction<Iterator<T>>) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- foreachPartitionAsync(VoidFunction<Iterator<T>>) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- foreachPartitionAsync(VoidFunction<Iterator<T>>) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
The asynchronous version of the foreachPartition action, which
applies a function f to each partition of this RDD.
- foreachPartitionAsync(Function1<Iterator<T>, BoxedUnit>) - 类 中的方法org.apache.spark.rdd.AsyncRDDActions
-
Applies a function f to each partition of this RDD.
- ForeachPartitionFunction<T> - org.apache.spark.api.java.function中的接口
-
Base interface for a function used in Dataset's foreachPartition function.
- foreachRDD(VoidFunction<R>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- foreachRDD(VoidFunction2<R, Time>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- foreachRDD(VoidFunction<R>) - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
Apply a function to each RDD in this DStream.
- foreachRDD(VoidFunction2<R, Time>) - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
Apply a function to each RDD in this DStream.
- foreachRDD(VoidFunction<R>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- foreachRDD(VoidFunction2<R, Time>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- foreachRDD(VoidFunction<R>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- foreachRDD(VoidFunction2<R, Time>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- foreachRDD(VoidFunction<R>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- foreachRDD(VoidFunction2<R, Time>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- foreachRDD(VoidFunction<R>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- foreachRDD(VoidFunction2<R, Time>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- foreachRDD(VoidFunction<R>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- foreachRDD(VoidFunction2<R, Time>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- foreachRDD(Function1<RDD<T>, BoxedUnit>) - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Apply a function to each RDD in this DStream.
- foreachRDD(Function2<RDD<T>, Time, BoxedUnit>) - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Apply a function to each RDD in this DStream.
- foreachUp(Function1<BaseType, BoxedUnit>) - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- foreachUp(Function1<BaseType, BoxedUnit>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- foreachUp(Function1<BaseType, BoxedUnit>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- foreachUp(Function1<BaseType, BoxedUnit>) - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- ForeachWriter<T> - org.apache.spark.sql中的类
-
A class to consume data generated by a StreamingQuery.
- ForeachWriter() - 类 的构造器org.apache.spark.sql.ForeachWriter
-
- format(String) - 类 中的方法org.apache.spark.sql.DataFrameReader
-
Specifies the input data source format.
- format(String) - 类 中的方法org.apache.spark.sql.DataFrameWriter
-
Specifies the underlying output data source.
- format(String) - 类 中的方法org.apache.spark.sql.streaming.DataStreamReader
-
Specifies the input data source format.
- format(String) - 类 中的方法org.apache.spark.sql.streaming.DataStreamWriter
-
Specifies the underlying output data source.
- format_number(Column, int) - 类 中的静态方法org.apache.spark.sql.functions
-
Formats numeric column x to a format like '#,###,###.##', rounded to d decimal places
with HALF_EVEN round mode, and returns the result as a string column.
- format_string(String, Column...) - 类 中的静态方法org.apache.spark.sql.functions
-
Formats the arguments in printf-style and returns the result as a string column.
- format_string(String, Seq<Column>) - 类 中的静态方法org.apache.spark.sql.functions
-
Formats the arguments in printf-style and returns the result as a string column.
- formatBatchTime(long, long, boolean, TimeZone) - 类 中的静态方法org.apache.spark.streaming.ui.UIUtils
-
If batchInterval is less than 1 second, format batchTime with milliseconds.
- formatDate(Date) - 类 中的静态方法org.apache.spark.ui.UIUtils
-
- formatDate(long) - 类 中的静态方法org.apache.spark.ui.UIUtils
-
- formatDuration(long) - 类 中的静态方法org.apache.spark.ui.UIUtils
-
- formatDurationVerbose(long) - 类 中的静态方法org.apache.spark.ui.UIUtils
-
Generate a verbose human-readable string representing a duration such as "5 second 35 ms"
- formatNumber(double) - 类 中的静态方法org.apache.spark.ui.UIUtils
-
Generate a human-readable string representing a number (e.g. 100 K)
- formatVersion() - 接口 中的方法org.apache.spark.mllib.util.Saveable
-
Current version of model save/load format.
- formula() - 类 中的静态方法org.apache.spark.ml.feature.RFormula
-
- formula() - 类 中的静态方法org.apache.spark.ml.feature.RFormulaModel
-
- FPGrowth - org.apache.spark.ml.fpm中的类
-
:: Experimental ::
A parallel FP-growth algorithm to mine frequent itemsets.
- FPGrowth(String) - 类 的构造器org.apache.spark.ml.fpm.FPGrowth
-
- FPGrowth() - 类 的构造器org.apache.spark.ml.fpm.FPGrowth
-
- FPGrowth - org.apache.spark.mllib.fpm中的类
-
A parallel FP-growth algorithm to mine frequent itemsets.
- FPGrowth() - 类 的构造器org.apache.spark.mllib.fpm.FPGrowth
-
Constructs a default instance with default parameters {minSupport: 0.3, numPartitions: same
as the input data}.
- FPGrowth.FreqItemset<Item> - org.apache.spark.mllib.fpm中的类
-
Frequent itemset.
- FPGrowthModel - org.apache.spark.ml.fpm中的类
-
:: Experimental ::
Model fitted by FPGrowth.
- FPGrowthModel<Item> - org.apache.spark.mllib.fpm中的类
-
Model trained by
FPGrowth, which holds frequent itemsets.
- FPGrowthModel(RDD<FPGrowth.FreqItemset<Item>>, ClassTag<Item>) - 类 的构造器org.apache.spark.mllib.fpm.FPGrowthModel
-
- FPGrowthModel.SaveLoadV1_0$ - org.apache.spark.mllib.fpm中的类
-
- fpr() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelector
-
- fpr() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- fpr() - 类 中的方法org.apache.spark.mllib.feature.ChiSqSelector
-
- freq() - 类 中的方法org.apache.spark.mllib.fpm.FPGrowth.FreqItemset
-
- freq() - 类 中的方法org.apache.spark.mllib.fpm.PrefixSpan.FreqSequence
-
- freqItems(String[], double) - 类 中的方法org.apache.spark.sql.DataFrameStatFunctions
-
Finding frequent items for columns, possibly with false positives.
- freqItems(String[]) - 类 中的方法org.apache.spark.sql.DataFrameStatFunctions
-
Finding frequent items for columns, possibly with false positives.
- freqItems(Seq<String>, double) - 类 中的方法org.apache.spark.sql.DataFrameStatFunctions
-
(Scala-specific) Finding frequent items for columns, possibly with false positives.
- freqItems(Seq<String>) - 类 中的方法org.apache.spark.sql.DataFrameStatFunctions
-
(Scala-specific) Finding frequent items for columns, possibly with false positives.
- FreqItemset(Object, long) - 类 的构造器org.apache.spark.mllib.fpm.FPGrowth.FreqItemset
-
- freqItemsets() - 类 中的方法org.apache.spark.ml.fpm.FPGrowthModel
-
- freqItemsets() - 类 中的方法org.apache.spark.mllib.fpm.FPGrowthModel
-
- FreqSequence(Object[], long) - 类 的构造器org.apache.spark.mllib.fpm.PrefixSpan.FreqSequence
-
- freqSequences() - 类 中的方法org.apache.spark.mllib.fpm.PrefixSpanModel
-
- from_json(Column, StructType, Map<String, String>) - 类 中的静态方法org.apache.spark.sql.functions
-
(Scala-specific) Parses a column containing a JSON string into a StructType with the
specified schema.
- from_json(Column, DataType, Map<String, String>) - 类 中的静态方法org.apache.spark.sql.functions
-
(Scala-specific) Parses a column containing a JSON string into a StructType or ArrayType
of StructTypes with the specified schema.
- from_json(Column, StructType, Map<String, String>) - 类 中的静态方法org.apache.spark.sql.functions
-
(Java-specific) Parses a column containing a JSON string into a StructType with the
specified schema.
- from_json(Column, DataType, Map<String, String>) - 类 中的静态方法org.apache.spark.sql.functions
-
(Java-specific) Parses a column containing a JSON string into a StructType or ArrayType
of StructTypes with the specified schema.
- from_json(Column, StructType) - 类 中的静态方法org.apache.spark.sql.functions
-
Parses a column containing a JSON string into a StructType with the specified schema.
- from_json(Column, DataType) - 类 中的静态方法org.apache.spark.sql.functions
-
Parses a column containing a JSON string into a StructType or ArrayType of StructTypes
with the specified schema.
- from_json(Column, String, Map<String, String>) - 类 中的静态方法org.apache.spark.sql.functions
-
(Java-specific) Parses a column containing a JSON string into a StructType or ArrayType
of StructTypes with the specified schema.
- from_json(Column, String, Map<String, String>) - 类 中的静态方法org.apache.spark.sql.functions
-
(Scala-specific) Parses a column containing a JSON string into a StructType or ArrayType
of StructTypes with the specified schema.
- from_unixtime(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Converts the number of seconds from unix epoch (1970-01-01 00:00:00 UTC) to a string
representing the timestamp of that moment in the current system time zone in the given
format.
- from_unixtime(Column, String) - 类 中的静态方法org.apache.spark.sql.functions
-
Converts the number of seconds from unix epoch (1970-01-01 00:00:00 UTC) to a string
representing the timestamp of that moment in the current system time zone in the given
format.
- from_utc_timestamp(Column, String) - 类 中的静态方法org.apache.spark.sql.functions
-
Given a timestamp like '2017-07-14 02:40:00.0', interprets it as a time in UTC, and renders
that time as a timestamp in the given time zone.
- fromCOO(int, int, Iterable<Tuple3<Object, Object, Object>>) - 类 中的静态方法org.apache.spark.ml.linalg.SparseMatrix
-
Generate a SparseMatrix from Coordinate List (COO) format.
- fromCOO(int, int, Iterable<Tuple3<Object, Object, Object>>) - 类 中的静态方法org.apache.spark.mllib.linalg.SparseMatrix
-
Generate a SparseMatrix from Coordinate List (COO) format.
- fromDDL(String) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
Creates StructType for a given DDL-formatted string, which is a comma separated list of field
definitions, e.g., a INT, b STRING.
- fromDecimal(Object) - 类 中的静态方法org.apache.spark.sql.types.Decimal
-
- fromDStream(DStream<T>, ClassTag<T>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- fromEdgePartitions(RDD<Tuple2<Object, EdgePartition<ED, VD>>>, VD, StorageLevel, StorageLevel, ClassTag<VD>, ClassTag<ED>) - 类 中的静态方法org.apache.spark.graphx.impl.GraphImpl
-
Create a graph from EdgePartitions, setting referenced vertices to defaultVertexAttr.
- fromEdges(RDD<Edge<ED>>, ClassTag<ED>, ClassTag<VD>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
Creates an EdgeRDD from a set of edges.
- fromEdges(RDD<Edge<ED>>, VD, StorageLevel, StorageLevel, ClassTag<VD>, ClassTag<ED>) - 类 中的静态方法org.apache.spark.graphx.Graph
-
Construct a graph from a collection of edges.
- fromEdges(EdgeRDD<?>, int, VD, ClassTag<VD>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
Constructs a VertexRDD containing all vertices referred to in edges.
- fromEdgeTuples(RDD<Tuple2<Object, Object>>, VD, Option<PartitionStrategy>, StorageLevel, StorageLevel, ClassTag<VD>) - 类 中的静态方法org.apache.spark.graphx.Graph
-
Construct a graph from a collection of edges encoded as vertex id pairs.
- fromExistingRDDs(VertexRDD<VD>, EdgeRDD<ED>, ClassTag<VD>, ClassTag<ED>) - 类 中的静态方法org.apache.spark.graphx.impl.GraphImpl
-
Create a graph from a VertexRDD and an EdgeRDD with the same replicated vertex type as the
vertices.
- fromInputDStream(InputDStream<T>, ClassTag<T>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- fromInputDStream(InputDStream<Tuple2<K, V>>, ClassTag<K>, ClassTag<V>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- fromJavaDStream(JavaDStream<Tuple2<K, V>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- fromJavaRDD(JavaRDD<Tuple2<K, V>>) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
Convert a JavaRDD of key-value pairs to JavaPairRDD.
- fromJson(String) - 类 中的静态方法org.apache.spark.ml.linalg.JsonMatrixConverter
-
Parses the JSON representation of a Matrix into a
Matrix.
- fromJson(String) - 类 中的静态方法org.apache.spark.ml.linalg.JsonVectorConverter
-
Parses the JSON representation of a vector into a
Vector.
- fromJson(String) - 类 中的静态方法org.apache.spark.mllib.linalg.Vectors
-
Parses the JSON representation of a vector into a
Vector.
- fromJson(String) - 类 中的静态方法org.apache.spark.sql.types.DataType
-
- fromJson(String) - 类 中的静态方法org.apache.spark.sql.types.Metadata
-
Creates a Metadata instance from JSON.
- fromKinesisInitialPosition(InitialPositionInStream) - 类 中的静态方法org.apache.spark.streaming.kinesis.KinesisInitialPositions
-
Returns instance of [[KinesisInitialPosition]] based on the passed
[[InitialPositionInStream]].
- fromML(DenseMatrix) - 类 中的静态方法org.apache.spark.mllib.linalg.DenseMatrix
-
Convert new linalg type to spark.mllib type.
- fromML(DenseVector) - 类 中的静态方法org.apache.spark.mllib.linalg.DenseVector
-
Convert new linalg type to spark.mllib type.
- fromML(Matrix) - 类 中的静态方法org.apache.spark.mllib.linalg.Matrices
-
Convert new linalg type to spark.mllib type.
- fromML(SparseMatrix) - 类 中的静态方法org.apache.spark.mllib.linalg.SparseMatrix
-
Convert new linalg type to spark.mllib type.
- fromML(SparseVector) - 类 中的静态方法org.apache.spark.mllib.linalg.SparseVector
-
Convert new linalg type to spark.mllib type.
- fromML(Vector) - 类 中的静态方法org.apache.spark.mllib.linalg.Vectors
-
Convert new linalg type to spark.mllib type.
- fromName(String) - 类 中的静态方法org.apache.spark.ml.attribute.AttributeType
-
- fromNullable(T) - 类 中的静态方法org.apache.spark.api.java.Optional
-
- fromOld(Node, Map<Object, Object>) - 类 中的静态方法org.apache.spark.ml.tree.Node
-
Create a new Node from the old Node format, recursively creating child nodes as needed.
- fromPairDStream(DStream<Tuple2<K, V>>, ClassTag<K>, ClassTag<V>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- fromPairRDD(RDD<Tuple2<K, V>>, ClassTag<K>, ClassTag<V>) - 类 中的静态方法org.apache.spark.mllib.rdd.MLPairRDDFunctions
-
Implicit conversion from a pair RDD to MLPairRDDFunctions.
- fromParams(GeneralizedLinearRegressionBase) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.Family$
-
Gets the Family object based on param family and variancePower.
- fromParams(GeneralizedLinearRegressionBase) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.Link$
-
Gets the Link object based on param family, link and linkPower.
- fromRDD(RDD<Object>) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- fromRDD(RDD<Tuple2<K, V>>, ClassTag<K>, ClassTag<V>) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- fromRDD(RDD<T>, ClassTag<T>) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- fromRDD(RDD<T>, ClassTag<T>) - 类 中的静态方法org.apache.spark.mllib.rdd.RDDFunctions
-
Implicit conversion from an RDD to RDDFunctions.
- fromRdd(RDD<?>) - 类 中的静态方法org.apache.spark.storage.RDDInfo
-
- fromReceiverInputDStream(ReceiverInputDStream<Tuple2<K, V>>, ClassTag<K>, ClassTag<V>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- fromReceiverInputDStream(ReceiverInputDStream<T>, ClassTag<T>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- fromSparkContext(SparkContext) - 类 中的静态方法org.apache.spark.api.java.JavaSparkContext
-
- fromStage(Stage, int, Option<Object>, TaskMetrics, Seq<Seq<TaskLocation>>) - 类 中的静态方法org.apache.spark.scheduler.StageInfo
-
Construct a StageInfo from a Stage.
- fromString(String) - 枚举 中的静态方法org.apache.spark.JobExecutionStatus
-
- fromString(String) - 类 中的静态方法org.apache.spark.mllib.tree.impurity.Impurities
-
- fromString(String) - 类 中的静态方法org.apache.spark.mllib.tree.loss.Losses
-
- fromString(String) - 枚举 中的静态方法org.apache.spark.status.api.v1.ApplicationStatus
-
- fromString(String) - 枚举 中的静态方法org.apache.spark.status.api.v1.StageStatus
-
- fromString(String) - 枚举 中的静态方法org.apache.spark.status.api.v1.streaming.BatchStatus
-
- fromString(String) - 枚举 中的静态方法org.apache.spark.status.api.v1.TaskSorting
-
- fromString(String) - 类 中的静态方法org.apache.spark.storage.StorageLevel
-
:: DeveloperApi ::
Return the StorageLevel object with the specified name.
- fromStructField(StructField) - 类 中的静态方法org.apache.spark.ml.attribute.AttributeGroup
-
Creates an attribute group from a StructField instance.
- fullOuterJoin(JavaPairRDD<K, W>, Partitioner) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Perform a full outer join of this and other.
- fullOuterJoin(JavaPairRDD<K, W>) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Perform a full outer join of this and other.
- fullOuterJoin(JavaPairRDD<K, W>, int) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Perform a full outer join of this and other.
- fullOuterJoin(RDD<Tuple2<K, W>>, Partitioner) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Perform a full outer join of this and other.
- fullOuterJoin(RDD<Tuple2<K, W>>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Perform a full outer join of this and other.
- fullOuterJoin(RDD<Tuple2<K, W>>, int) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Perform a full outer join of this and other.
- fullOuterJoin(JavaPairDStream<K, W>) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying 'full outer join' between RDDs of this DStream and
other DStream.
- fullOuterJoin(JavaPairDStream<K, W>, int) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying 'full outer join' between RDDs of this DStream and
other DStream.
- fullOuterJoin(JavaPairDStream<K, W>, Partitioner) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying 'full outer join' between RDDs of this DStream and
other DStream.
- fullOuterJoin(JavaPairDStream<K, W>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- fullOuterJoin(JavaPairDStream<K, W>, int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- fullOuterJoin(JavaPairDStream<K, W>, Partitioner) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- fullOuterJoin(JavaPairDStream<K, W>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- fullOuterJoin(JavaPairDStream<K, W>, int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- fullOuterJoin(JavaPairDStream<K, W>, Partitioner) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- fullOuterJoin(DStream<Tuple2<K, W>>, ClassTag<W>) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying 'full outer join' between RDDs of this DStream and
other DStream.
- fullOuterJoin(DStream<Tuple2<K, W>>, int, ClassTag<W>) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying 'full outer join' between RDDs of this DStream and
other DStream.
- fullOuterJoin(DStream<Tuple2<K, W>>, Partitioner, ClassTag<W>) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying 'full outer join' between RDDs of this DStream and
other DStream.
- fullStackTrace() - 类 中的方法org.apache.spark.ExceptionFailure
-
- Function<T1,R> - org.apache.spark.api.java.function中的接口
-
Base interface for functions whose return types do not create special RDDs.
- Function - org.apache.spark.sql.catalog中的类
-
A user-defined function in Spark, as returned by
listFunctions method in
Catalog.
- Function(String, String, String, String, boolean) - 类 的构造器org.apache.spark.sql.catalog.Function
-
- function(Function4<Time, KeyType, Option<ValueType>, State<StateType>, Option<MappedType>>) - 类 中的静态方法org.apache.spark.streaming.StateSpec
-
- function(Function3<KeyType, Option<ValueType>, State<StateType>, MappedType>) - 类 中的静态方法org.apache.spark.streaming.StateSpec
-
- function(Function4<Time, KeyType, Optional<ValueType>, State<StateType>, Optional<MappedType>>) - 类 中的静态方法org.apache.spark.streaming.StateSpec
-
- function(Function3<KeyType, Optional<ValueType>, State<StateType>, MappedType>) - 类 中的静态方法org.apache.spark.streaming.StateSpec
-
- Function0<R> - org.apache.spark.api.java.function中的接口
-
A zero-argument function that returns an R.
- Function2<T1,T2,R> - org.apache.spark.api.java.function中的接口
-
A two-argument function that takes arguments of type T1 and T2 and returns an R.
- Function3<T1,T2,T3,R> - org.apache.spark.api.java.function中的接口
-
A three-argument function that takes arguments of type T1, T2 and T3 and returns an R.
- Function4<T1,T2,T3,T4,R> - org.apache.spark.api.java.function中的接口
-
A four-argument function that takes arguments of type T1, T2, T3 and T4 and returns an R.
- functionExists(String) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
Check if the function with the specified name exists.
- functionExists(String, String) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
Check if the function with the specified name exists in the specified database.
- functions - org.apache.spark.sql中的类
-
Functions available for DataFrame operations.
- functions() - 类 的构造器org.apache.spark.sql.functions
-
- FutureAction<T> - org.apache.spark中的接口
-
A future for the result of an action to support cancellation.
- futureExecutionContext() - 类 中的静态方法org.apache.spark.rdd.AsyncRDDActions
-
- fwe() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelector
-
- fwe() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- fwe() - 类 中的方法org.apache.spark.mllib.feature.ChiSqSelector
-
- gain() - 类 中的方法org.apache.spark.ml.tree.DecisionTreeModelReadWrite.NodeData
-
- gain() - 类 中的方法org.apache.spark.ml.tree.InternalNode
-
- gain() - 类 中的方法org.apache.spark.mllib.tree.model.InformationGainStats
-
- Gamma$() - 类 的构造器org.apache.spark.ml.regression.GeneralizedLinearRegression.Gamma$
-
- gamma1() - 类 中的方法org.apache.spark.graphx.lib.SVDPlusPlus.Conf
-
- gamma2() - 类 中的方法org.apache.spark.graphx.lib.SVDPlusPlus.Conf
-
- gamma6() - 类 中的方法org.apache.spark.graphx.lib.SVDPlusPlus.Conf
-
- gamma7() - 类 中的方法org.apache.spark.graphx.lib.SVDPlusPlus.Conf
-
- GammaGenerator - org.apache.spark.mllib.random中的类
-
:: DeveloperApi ::
Generates i.i.d. samples from the gamma distribution with the given shape and scale.
- GammaGenerator(double, double) - 类 的构造器org.apache.spark.mllib.random.GammaGenerator
-
- gammaJavaRDD(JavaSparkContext, double, double, long, int, long) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
Java-friendly version of RandomRDDs.gammaRDD.
- gammaJavaRDD(JavaSparkContext, double, double, long, int) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
RandomRDDs.gammaJavaRDD with the default seed.
- gammaJavaRDD(JavaSparkContext, double, double, long) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
RandomRDDs.gammaJavaRDD with the default number of partitions and the default seed.
- gammaJavaVectorRDD(JavaSparkContext, double, double, long, int, int, long) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
Java-friendly version of RandomRDDs.gammaVectorRDD.
- gammaJavaVectorRDD(JavaSparkContext, double, double, long, int, int) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
RandomRDDs.gammaJavaVectorRDD with the default seed.
- gammaJavaVectorRDD(JavaSparkContext, double, double, long, int) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
RandomRDDs.gammaJavaVectorRDD with the default number of partitions and the default seed.
- gammaRDD(SparkContext, double, double, long, int, long) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
Generates an RDD comprised of i.i.d.
- gammaVectorRDD(SparkContext, double, double, long, int, int, long) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
Generates an RDD[Vector] with vectors containing i.i.d.
- gapply(RelationalGroupedDataset, byte[], byte[], Object[], StructType) - 类 中的静态方法org.apache.spark.sql.api.r.SQLUtils
-
The helper function for gapply() on R side.
- gaps() - 类 中的方法org.apache.spark.ml.feature.RegexTokenizer
-
Indicates whether regex splits on gaps (true) or matches tokens (false).
- GAUGE() - 类 中的静态方法org.apache.spark.metrics.sink.StatsdMetricType
-
- Gaussian$() - 类 的构造器org.apache.spark.ml.regression.GeneralizedLinearRegression.Gaussian$
-
- GaussianMixture - org.apache.spark.ml.clustering中的类
-
Gaussian Mixture clustering.
- GaussianMixture(String) - 类 的构造器org.apache.spark.ml.clustering.GaussianMixture
-
- GaussianMixture() - 类 的构造器org.apache.spark.ml.clustering.GaussianMixture
-
- GaussianMixture - org.apache.spark.mllib.clustering中的类
-
This class performs expectation maximization for multivariate Gaussian
Mixture Models (GMMs).
- GaussianMixture() - 类 的构造器org.apache.spark.mllib.clustering.GaussianMixture
-
Constructs a default instance.
- GaussianMixtureModel - org.apache.spark.ml.clustering中的类
-
Multivariate Gaussian Mixture Model (GMM) consisting of k Gaussians, where points
are drawn from each Gaussian i with probability weights(i).
- GaussianMixtureModel - org.apache.spark.mllib.clustering中的类
-
Multivariate Gaussian Mixture Model (GMM) consisting of k Gaussians, where points
are drawn from each Gaussian i=1..k with probability w(i); mu(i) and sigma(i) are
the respective mean and covariance for each Gaussian distribution i=1..k.
- GaussianMixtureModel(double[], MultivariateGaussian[]) - 类 的构造器org.apache.spark.mllib.clustering.GaussianMixtureModel
-
- GaussianMixtureSummary - org.apache.spark.ml.clustering中的类
-
:: Experimental ::
Summary of GaussianMixture.
- gaussians() - 类 中的方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- gaussians() - 类 中的方法org.apache.spark.mllib.clustering.GaussianMixtureModel
-
- gaussiansDF() - 类 中的方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
Retrieve Gaussian distributions as a DataFrame.
- GBTClassificationModel - org.apache.spark.ml.classification中的类
-
Gradient-Boosted Trees (GBTs) (http://en.wikipedia.org/wiki/Gradient_boosting)
model for classification.
- GBTClassificationModel(String, DecisionTreeRegressionModel[], double[]) - 类 的构造器org.apache.spark.ml.classification.GBTClassificationModel
-
Construct a GBTClassificationModel
- GBTClassifier - org.apache.spark.ml.classification中的类
-
Gradient-Boosted Trees (GBTs) (http://en.wikipedia.org/wiki/Gradient_boosting)
learning algorithm for classification.
- GBTClassifier(String) - 类 的构造器org.apache.spark.ml.classification.GBTClassifier
-
- GBTClassifier() - 类 的构造器org.apache.spark.ml.classification.GBTClassifier
-
- GBTRegressionModel - org.apache.spark.ml.regression中的类
-
- GBTRegressionModel(String, DecisionTreeRegressionModel[], double[]) - 类 的构造器org.apache.spark.ml.regression.GBTRegressionModel
-
Construct a GBTRegressionModel
- GBTRegressor - org.apache.spark.ml.regression中的类
-
- GBTRegressor(String) - 类 的构造器org.apache.spark.ml.regression.GBTRegressor
-
- GBTRegressor() - 类 的构造器org.apache.spark.ml.regression.GBTRegressor
-
- GC_TIME() - 类 中的静态方法org.apache.spark.status.TaskIndexNames
-
- GC_TIME() - 类 中的静态方法org.apache.spark.ui.ToolTips
-
- gemm(double, Matrix, DenseMatrix, double, DenseMatrix) - 类 中的静态方法org.apache.spark.ml.linalg.BLAS
-
C := alpha * A * B + beta * C
- gemm(double, Matrix, DenseMatrix, double, DenseMatrix) - 类 中的静态方法org.apache.spark.mllib.linalg.BLAS
-
C := alpha * A * B + beta * C
- gemv(double, Matrix, Vector, double, DenseVector) - 类 中的静态方法org.apache.spark.ml.linalg.BLAS
-
y := alpha * A * x + beta * y
- gemv(double, Matrix, Vector, double, DenseVector) - 类 中的静态方法org.apache.spark.mllib.linalg.BLAS
-
y := alpha * A * x + beta * y
- GeneralizedLinearAlgorithm<M extends GeneralizedLinearModel> - org.apache.spark.mllib.regression中的类
-
:: DeveloperApi ::
GeneralizedLinearAlgorithm implements methods to train a Generalized Linear Model (GLM).
- GeneralizedLinearAlgorithm() - 类 的构造器org.apache.spark.mllib.regression.GeneralizedLinearAlgorithm
-
- GeneralizedLinearModel - org.apache.spark.mllib.regression中的类
-
:: DeveloperApi ::
GeneralizedLinearModel (GLM) represents a model trained using
GeneralizedLinearAlgorithm.
- GeneralizedLinearModel(Vector, double) - 类 的构造器org.apache.spark.mllib.regression.GeneralizedLinearModel
-
- GeneralizedLinearRegression - org.apache.spark.ml.regression中的类
-
:: Experimental ::
Fit a Generalized Linear Model
(see
Generalized linear model (Wikipedia))
specified by giving a symbolic description of the linear
predictor (link function) and a description of the error distribution (family).
- GeneralizedLinearRegression(String) - 类 的构造器org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- GeneralizedLinearRegression() - 类 的构造器org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- GeneralizedLinearRegression.Binomial$ - org.apache.spark.ml.regression中的类
-
Binomial exponential family distribution.
- GeneralizedLinearRegression.CLogLog$ - org.apache.spark.ml.regression中的类
-
- GeneralizedLinearRegression.Family$ - org.apache.spark.ml.regression中的类
-
- GeneralizedLinearRegression.FamilyAndLink$ - org.apache.spark.ml.regression中的类
-
- GeneralizedLinearRegression.Gamma$ - org.apache.spark.ml.regression中的类
-
Gamma exponential family distribution.
- GeneralizedLinearRegression.Gaussian$ - org.apache.spark.ml.regression中的类
-
Gaussian exponential family distribution.
- GeneralizedLinearRegression.Identity$ - org.apache.spark.ml.regression中的类
-
- GeneralizedLinearRegression.Inverse$ - org.apache.spark.ml.regression中的类
-
- GeneralizedLinearRegression.Link$ - org.apache.spark.ml.regression中的类
-
- GeneralizedLinearRegression.Log$ - org.apache.spark.ml.regression中的类
-
- GeneralizedLinearRegression.Logit$ - org.apache.spark.ml.regression中的类
-
- GeneralizedLinearRegression.Poisson$ - org.apache.spark.ml.regression中的类
-
Poisson exponential family distribution.
- GeneralizedLinearRegression.Probit$ - org.apache.spark.ml.regression中的类
-
- GeneralizedLinearRegression.Sqrt$ - org.apache.spark.ml.regression中的类
-
- GeneralizedLinearRegression.Tweedie$ - org.apache.spark.ml.regression中的类
-
- GeneralizedLinearRegressionModel - org.apache.spark.ml.regression中的类
-
- GeneralizedLinearRegressionSummary - org.apache.spark.ml.regression中的类
-
- GeneralizedLinearRegressionTrainingSummary - org.apache.spark.ml.regression中的类
-
- generateAssociationRules(double) - 类 中的方法org.apache.spark.mllib.fpm.FPGrowthModel
-
Generates association rules for the Items in freqItemsets.
- generateKMeansRDD(SparkContext, int, int, int, double, int) - 类 中的静态方法org.apache.spark.mllib.util.KMeansDataGenerator
-
Generate an RDD containing test data for KMeans.
- generateLinearInput(double, double[], int, int, double) - 类 中的静态方法org.apache.spark.mllib.util.LinearDataGenerator
-
For compatibility, the generated data without specifying the mean and variance
will have zero mean and variance of (1.0/3.0) since the original output range is
[-1, 1] with uniform distribution, and the variance of uniform distribution
is (b - a)^2^ / 12 which will be (1.0/3.0)
- generateLinearInput(double, double[], double[], double[], int, int, double) - 类 中的静态方法org.apache.spark.mllib.util.LinearDataGenerator
-
- generateLinearInput(double, double[], double[], double[], int, int, double, double) - 类 中的静态方法org.apache.spark.mllib.util.LinearDataGenerator
-
- generateLinearInputAsList(double, double[], int, int, double) - 类 中的静态方法org.apache.spark.mllib.util.LinearDataGenerator
-
Return a Java List of synthetic data randomly generated according to a multi
collinear model.
- generateLinearRDD(SparkContext, int, int, double, int, double) - 类 中的静态方法org.apache.spark.mllib.util.LinearDataGenerator
-
Generate an RDD containing sample data for Linear Regression models - including Ridge, Lasso,
and unregularized variants.
- generateLogisticRDD(SparkContext, int, int, double, int, double) - 类 中的静态方法org.apache.spark.mllib.util.LogisticRegressionDataGenerator
-
Generate an RDD containing test data for LogisticRegression.
- generateRandomEdges(int, int, int, long) - 类 中的静态方法org.apache.spark.graphx.util.GraphGenerators
-
- generateTreeString(int, Seq<Object>, StringBuilder, boolean, String, boolean) - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- generateTreeString(int, Seq<Object>, StringBuilder, boolean, String, boolean) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- generateTreeString(int, Seq<Object>, StringBuilder, boolean, String, boolean) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- generateTreeString(int, Seq<Object>, StringBuilder, boolean, String, boolean) - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- generateTreeString$default$5() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- generateTreeString$default$5() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- generateTreeString$default$5() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- generateTreeString$default$5() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- generateTreeString$default$6() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- generateTreeString$default$6() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- generateTreeString$default$6() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- generateTreeString$default$6() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- genericBuilder() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- geq(Object) - 类 中的方法org.apache.spark.sql.Column
-
Greater than or equal to an expression.
- get(Object) - 类 中的方法org.apache.spark.api.java.JavaUtils.SerializableMapWrapper
-
- get() - 类 中的方法org.apache.spark.api.java.Optional
-
- get() - 接口 中的方法org.apache.spark.FutureAction
-
Blocks and returns the result of this job.
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.OneVsRest
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.OneVsRestModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeans
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixture
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.clustering.KMeans
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.clustering.KMeansModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.evaluation.ClusteringEvaluator
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.evaluation.RegressionEvaluator
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.Binarizer
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.Bucketizer
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelector
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.ColumnPruner
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizer
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizerModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.DCT
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.ElementwiseProduct
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.FeatureHasher
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.HashingTF
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.IDF
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.IDFModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.Imputer
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.ImputerModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.IndexToString
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.Interaction
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.MaxAbsScaler
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.MaxAbsScalerModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSH
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSHModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScaler
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScalerModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.NGram
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.Normalizer
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoder
-
已过时。
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderEstimator
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.PCA
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.PCAModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.PolynomialExpansion
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.QuantileDiscretizer
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.RegexTokenizer
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.RFormula
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.SQLTransformer
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.StandardScaler
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.StandardScalerModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.StopWordsRemover
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.StringIndexer
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.StringIndexerModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.Tokenizer
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.VectorAssembler
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.VectorAttributeRewriter
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexer
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexerModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.VectorSizeHint
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.VectorSlicer
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.Word2Vec
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.Word2VecModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowth
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowthModel
-
- get(Param<T>) - 类 中的方法org.apache.spark.ml.param.ParamMap
-
Optionally returns the value associated with a param.
- get(Param<T>) - 接口 中的方法org.apache.spark.ml.param.Params
-
Optionally returns the user-supplied value of a param.
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.Pipeline
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.PipelineModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.recommendation.ALSModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegression
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegressionModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidator
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidatorModel
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplit
-
- get(Param<T>) - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplitModel
-
- get(String) - 类 中的方法org.apache.spark.SparkConf
-
Get a parameter; throws a NoSuchElementException if it's not set
- get(String, String) - 类 中的方法org.apache.spark.SparkConf
-
Get a parameter, falling back to a default if not set
- get() - 类 中的静态方法org.apache.spark.SparkEnv
-
Returns the SparkEnv.
- get(String) - 类 中的静态方法org.apache.spark.SparkFiles
-
Get the absolute path of a file added through SparkContext.addFile().
- get(String) - 类 中的静态方法org.apache.spark.sql.jdbc.JdbcDialects
-
Fetch the JdbcDialect class corresponding to a given database url.
- get(int) - 接口 中的方法org.apache.spark.sql.Row
-
Returns the value at position i.
- get(String) - 类 中的方法org.apache.spark.sql.RuntimeConfig
-
Returns the value of Spark runtime configuration property for the given key.
- get(String, String) - 类 中的方法org.apache.spark.sql.RuntimeConfig
-
Returns the value of Spark runtime configuration property for the given key.
- get(String) - 类 中的方法org.apache.spark.sql.sources.v2.DataSourceOptions
-
Returns the option value to which the specified key is mapped, case-insensitively.
- get() - 接口 中的方法org.apache.spark.sql.sources.v2.reader.DataReader
-
Return the current record.
- get() - 接口 中的方法org.apache.spark.sql.streaming.GroupState
-
Get the state value if it exists, or throw NoSuchElementException.
- get(UUID) - 类 中的方法org.apache.spark.sql.streaming.StreamingQueryManager
-
Returns the query if there is an active query with the given id, or null.
- get(String) - 类 中的方法org.apache.spark.sql.streaming.StreamingQueryManager
-
Returns the query if there is an active query with the given id, or null.
- get(int, DataType) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarArray
-
- get(int, DataType) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarRow
-
- get() - 类 中的方法org.apache.spark.streaming.State
-
Get the state if it exists, otherwise it will throw java.util.NoSuchElementException.
- get() - 类 中的静态方法org.apache.spark.TaskContext
-
Return the currently active TaskContext.
- get(long) - 类 中的静态方法org.apache.spark.util.AccumulatorContext
-
- get_json_object(Column, String) - 类 中的静态方法org.apache.spark.sql.functions
-
Extracts json object from a json string based on json path specified, and returns json string
of the extracted json object.
- getAcceptanceResults(RDD<Tuple2<K, V>>, boolean, Map<K, Object>, Option<Map<K, Object>>, long) - 类 中的静态方法org.apache.spark.util.random.StratifiedSamplingUtils
-
Count the number of items instantly accepted and generate the waitlist for each stratum.
- getAcceptsNull() - 类 中的静态方法org.apache.spark.serializer.JavaIterableWrapperSerializer
-
- getActive() - 类 中的静态方法org.apache.spark.streaming.StreamingContext
-
:: Experimental ::
Get the currently active context, if there is one.
- getActiveJobIds() - 类 中的方法org.apache.spark.api.java.JavaSparkStatusTracker
-
Returns an array containing the ids of all active jobs.
- getActiveJobIds() - 类 中的方法org.apache.spark.SparkStatusTracker
-
Returns an array containing the ids of all active jobs.
- getActiveOrCreate(Function0<StreamingContext>) - 类 中的静态方法org.apache.spark.streaming.StreamingContext
-
:: Experimental ::
Either return the "active" StreamingContext (that is, started but not stopped), or create a
new StreamingContext that is
- getActiveOrCreate(String, Function0<StreamingContext>, Configuration, boolean) - 类 中的静态方法org.apache.spark.streaming.StreamingContext
-
:: Experimental ::
Either get the currently active StreamingContext (that is, started but not stopped),
OR recreate a StreamingContext from checkpoint data in the given path.
- getActiveSession() - 类 中的静态方法org.apache.spark.sql.SparkSession
-
Returns the active SparkSession for the current thread, returned by the builder.
- getActiveStageIds() - 类 中的方法org.apache.spark.api.java.JavaSparkStatusTracker
-
Returns an array containing the ids of all active stages.
- getActiveStageIds() - 类 中的方法org.apache.spark.SparkStatusTracker
-
Returns an array containing the ids of all active stages.
- getAggregationDepth() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- getAggregationDepth() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- getAggregationDepth() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- getAggregationDepth() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- getAggregationDepth() - 接口 中的方法org.apache.spark.ml.param.shared.HasAggregationDepth
-
- getAggregationDepth() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- getAggregationDepth() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- getAggregationDepth() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- getAggregationDepth() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- getAlgo() - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- getAll() - 类 中的方法org.apache.spark.SparkConf
-
Get all parameters as a list of pairs
- getAll() - 类 中的方法org.apache.spark.sql.RuntimeConfig
-
Returns all properties set in this conf.
- getAllConfs() - 类 中的方法org.apache.spark.sql.SQLContext
-
Return all the configuration properties that have been set (i.e. not the default).
- getAllPools() - 类 中的方法org.apache.spark.SparkContext
-
:: DeveloperApi ::
Return pools for fair scheduler
- getAllPrefLocs(RDD<?>) - 类 中的方法org.apache.spark.rdd.DefaultPartitionCoalescer.PartitionLocations
-
- GetAllReceiverInfo - org.apache.spark.streaming.scheduler中的类
-
- GetAllReceiverInfo() - 类 的构造器org.apache.spark.streaming.scheduler.GetAllReceiverInfo
-
- getAllWithPrefix(String) - 类 中的方法org.apache.spark.SparkConf
-
Get all parameters that start with prefix
- getAlpha() - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- getAlpha() - 类 中的方法org.apache.spark.mllib.clustering.LDA
-
Alias for getDocConcentration
- getAnyValAs(int) - 接口 中的方法org.apache.spark.sql.Row
-
Returns the value at position i.
- getAppId() - 接口 中的方法org.apache.spark.launcher.SparkAppHandle
-
Returns the application ID, or null if not yet known.
- getAppId() - 类 中的方法org.apache.spark.SparkConf
-
Returns the Spark application id, valid in the Driver after TaskScheduler registration and
from the start in the Executor.
- getArray(int) - 类 中的方法org.apache.spark.sql.vectorized.ArrowColumnVector
-
- getArray(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarArray
-
- getArray(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarRow
-
- getArray(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnVector
-
Returns the array type value for rowId.
- getAs(int) - 接口 中的方法org.apache.spark.sql.Row
-
Returns the value at position i.
- getAs(String) - 接口 中的方法org.apache.spark.sql.Row
-
Returns the value of a given fieldName.
- getAssociationRulesFromFP(Dataset<?>, String, String, double, ClassTag<T>) - 类 中的静态方法org.apache.spark.ml.fpm.AssociationRules
-
Computes the association rules with confidence above minConfidence.
- getAsymmetricAlpha() - 类 中的方法org.apache.spark.mllib.clustering.LDA
-
Alias for getAsymmetricDocConcentration
- getAsymmetricDocConcentration() - 类 中的方法org.apache.spark.mllib.clustering.LDA
-
Concentration parameter (commonly named "alpha") for the prior placed on documents'
distributions over topics ("theta").
- getAttr(String) - 类 中的方法org.apache.spark.ml.attribute.AttributeGroup
-
Gets an attribute by its name.
- getAttr(int) - 类 中的方法org.apache.spark.ml.attribute.AttributeGroup
-
Gets an attribute by its index.
- getAvroSchema() - 类 中的方法org.apache.spark.SparkConf
-
Gets all the avro schemas in the configuration used in the generic Avro record serializer
- getBatchingTimeout(SparkConf) - 类 中的静态方法org.apache.spark.streaming.util.WriteAheadLogUtils
-
How long we will wait for the wrappedLog in the BatchedWriteAheadLog to write the records
before we fail the write attempt to unblock receivers.
- getBernoulliSamplingFunction(RDD<Tuple2<K, V>>, Map<K, Object>, boolean, long) - 类 中的静态方法org.apache.spark.util.random.StratifiedSamplingUtils
-
Return the per partition sampling function used for sampling without replacement.
- getBeta() - 类 中的方法org.apache.spark.mllib.clustering.LDA
-
Alias for getTopicConcentration
- getBinary() - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizer
-
- getBinary() - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizerModel
-
- getBinary() - 类 中的方法org.apache.spark.ml.feature.HashingTF
-
- getBinary(int) - 类 中的方法org.apache.spark.sql.vectorized.ArrowColumnVector
-
- getBinary(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarArray
-
- getBinary(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarRow
-
- getBinary(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnVector
-
Returns the binary type value for rowId.
- getBlock(BlockId) - 类 中的方法org.apache.spark.storage.StorageStatus
-
已过时。
Return the given block stored in this block manager in O(1) time.
- getBlockSize() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- GetBlockStatus(BlockId, boolean) - 类 的构造器org.apache.spark.storage.BlockManagerMessages.GetBlockStatus
-
- GetBlockStatus$() - 类 的构造器org.apache.spark.storage.BlockManagerMessages.GetBlockStatus$
-
- getBoolean(String, boolean) - 类 中的方法org.apache.spark.SparkConf
-
Get a parameter as a boolean, falling back to a default if not set
- getBoolean(int) - 接口 中的方法org.apache.spark.sql.Row
-
Returns the value at position i as a primitive boolean.
- getBoolean(String, boolean) - 类 中的方法org.apache.spark.sql.sources.v2.DataSourceOptions
-
Returns the boolean value to which the specified key is mapped,
or defaultValue if there is no mapping for the key.
- getBoolean(String) - 类 中的方法org.apache.spark.sql.types.Metadata
-
Gets a Boolean.
- getBoolean(int) - 类 中的方法org.apache.spark.sql.vectorized.ArrowColumnVector
-
- getBoolean(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarArray
-
- getBoolean(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarRow
-
- getBoolean(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnVector
-
Returns the boolean type value for rowId.
- getBooleanArray(String) - 类 中的方法org.apache.spark.sql.types.Metadata
-
Gets a Boolean array.
- getBooleans(int, int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnVector
-
Gets boolean type values from [rowId, rowId + count).
- getBucketLength() - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- getBucketLength() - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
-
- getByte(int) - 接口 中的方法org.apache.spark.sql.Row
-
Returns the value at position i as a primitive byte.
- getByte(int) - 类 中的方法org.apache.spark.sql.vectorized.ArrowColumnVector
-
- getByte(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarArray
-
- getByte(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarRow
-
- getByte(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnVector
-
Returns the byte type value for rowId.
- getBytes(int, int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnVector
-
Gets byte type values from [rowId, rowId + count).
- getCachedBlockManagerId(BlockManagerId) - 类 中的静态方法org.apache.spark.storage.BlockManagerId
-
- getCachedMetadata(String) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
The three methods below are helpers for accessing the local map, a property of the SparkEnv of
the local process.
- getCacheNodeIds() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- getCacheNodeIds() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- getCacheNodeIds() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- getCacheNodeIds() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- getCacheNodeIds() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- getCacheNodeIds() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- getCacheNodeIds() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- getCacheNodeIds() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- getCacheNodeIds() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- getCacheNodeIds() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- getCacheNodeIds() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- getCacheNodeIds() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- getCallSite(Function1<String, Object>) - 类 中的静态方法org.apache.spark.util.Utils
-
When called inside a class in the spark package, returns the name of the user code class
(outside the spark package) that called into Spark, as well as which Spark method they called.
- getCaseSensitive() - 类 中的方法org.apache.spark.ml.feature.StopWordsRemover
-
- getCatalystType(int, String, int, MetadataBuilder) - 类 中的方法org.apache.spark.sql.jdbc.AggregatedDialect
-
- getCatalystType(int, String, int, MetadataBuilder) - 类 中的静态方法org.apache.spark.sql.jdbc.DB2Dialect
-
- getCatalystType(int, String, int, MetadataBuilder) - 类 中的静态方法org.apache.spark.sql.jdbc.DerbyDialect
-
- getCatalystType(int, String, int, MetadataBuilder) - 类 中的方法org.apache.spark.sql.jdbc.JdbcDialect
-
Get the custom datatype mapping for the given jdbc meta information.
- getCatalystType(int, String, int, MetadataBuilder) - 类 中的静态方法org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- getCatalystType(int, String, int, MetadataBuilder) - 类 中的静态方法org.apache.spark.sql.jdbc.MySQLDialect
-
- getCatalystType(int, String, int, MetadataBuilder) - 类 中的静态方法org.apache.spark.sql.jdbc.NoopDialect
-
- getCatalystType(int, String, int, MetadataBuilder) - 类 中的静态方法org.apache.spark.sql.jdbc.OracleDialect
-
- getCatalystType(int, String, int, MetadataBuilder) - 类 中的静态方法org.apache.spark.sql.jdbc.PostgresDialect
-
- getCatalystType(int, String, int, MetadataBuilder) - 类 中的静态方法org.apache.spark.sql.jdbc.TeradataDialect
-
- getCategoricalCols() - 类 中的方法org.apache.spark.ml.feature.FeatureHasher
-
- getCategoricalFeatures(StructField) - 类 中的静态方法org.apache.spark.ml.util.MetadataUtils
-
Examine a schema to identify categorical (Binary and Nominal) features.
- getCategoricalFeaturesInfo() - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- getCause() - 异常错误 中的静态方法org.apache.spark.sql.AnalysisException
-
- getCensorCol() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- getCensorCol() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- getCheckpointDir() - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
- getCheckpointDir() - 类 中的方法org.apache.spark.SparkContext
-
- getCheckpointFile() - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- getCheckpointFile() - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- getCheckpointFile() - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- getCheckpointFile() - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Gets the name of the file to which this RDD was checkpointed
- getCheckpointFile() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- getCheckpointFile() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- getCheckpointFile() - 类 中的方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- getCheckpointFile() - 类 中的方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- getCheckpointFile() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- getCheckpointFile() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- getCheckpointFile() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- getCheckpointFile() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- getCheckpointFile() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- getCheckpointFile() - 类 中的方法org.apache.spark.rdd.RDD
-
Gets the name of the directory to which this RDD was checkpointed.
- getCheckpointFile() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- getCheckpointFiles() - 类 中的方法org.apache.spark.graphx.Graph
-
Gets the name of the files to which this Graph was checkpointed.
- getCheckpointFiles() - 类 中的方法org.apache.spark.graphx.impl.GraphImpl
-
- getCheckpointFiles() - 类 中的方法org.apache.spark.ml.clustering.DistributedLDAModel
-
:: DeveloperApi ::
If using checkpointing and LDA.keepLastCheckpoint is set to true, then there may be
saved checkpoint files.
- getCheckpointInterval() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- getCheckpointInterval() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- getCheckpointInterval() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- getCheckpointInterval() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- getCheckpointInterval() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- getCheckpointInterval() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- getCheckpointInterval() - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- getCheckpointInterval() - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- getCheckpointInterval() - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- getCheckpointInterval() - 接口 中的方法org.apache.spark.ml.param.shared.HasCheckpointInterval
-
- getCheckpointInterval() - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- getCheckpointInterval() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- getCheckpointInterval() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- getCheckpointInterval() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- getCheckpointInterval() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- getCheckpointInterval() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- getCheckpointInterval() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- getCheckpointInterval() - 类 中的方法org.apache.spark.mllib.clustering.LDA
-
Period (in iterations) between checkpoints.
- getCheckpointInterval() - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- getChild(int) - 类 中的方法org.apache.spark.sql.vectorized.ArrowColumnVector
-
- getClassifier() - 类 中的静态方法org.apache.spark.ml.classification.OneVsRest
-
- getClassifier() - 类 中的静态方法org.apache.spark.ml.classification.OneVsRestModel
-
- getColdStartStrategy() - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- getColdStartStrategy() - 类 中的静态方法org.apache.spark.ml.recommendation.ALSModel
-
- getCollectSubModels() - 接口 中的方法org.apache.spark.ml.param.shared.HasCollectSubModels
-
- getCollectSubModels() - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidator
-
- getCollectSubModels() - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplit
-
- getCombOp() - 类 中的静态方法org.apache.spark.util.random.StratifiedSamplingUtils
-
Returns the function used combine results returned by seqOp from different partitions.
- getComment() - 类 中的方法org.apache.spark.sql.types.StructField
-
Return the comment of this StructField.
- getConf() - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
Return a copy of this JavaSparkContext's configuration.
- getConf() - 类 中的静态方法org.apache.spark.ml.image.SamplePathFilter
-
- getConf() - 类 中的方法org.apache.spark.rdd.HadoopRDD
-
- getConf() - 类 中的方法org.apache.spark.rdd.NewHadoopRDD
-
- getConf() - 类 中的方法org.apache.spark.SparkContext
-
Return a copy of this SparkContext's configuration.
- getConf(String) - 类 中的方法org.apache.spark.sql.SQLContext
-
Return the value of Spark SQL configuration property for the given key.
- getConf(String, String) - 类 中的方法org.apache.spark.sql.SQLContext
-
Return the value of Spark SQL configuration property for the given key.
- getConfiguration() - 类 中的方法org.apache.spark.input.PortableDataStream
-
- getConfiguredLocalDirs(SparkConf) - 类 中的静态方法org.apache.spark.util.Utils
-
Return the configured local directories where Spark can write files.
- getConnection() - 接口 中的方法org.apache.spark.rdd.JdbcRDD.ConnectionFactory
-
- getContextOrSparkClassLoader() - 类 中的静态方法org.apache.spark.util.Utils
-
Get the Context ClassLoader on this thread or, if not present, the ClassLoader that
loaded Spark.
- getConvergenceTol() - 类 中的方法org.apache.spark.mllib.clustering.GaussianMixture
-
Return the largest change in log-likelihood at which convergence is
considered to have occurred.
- getCorrelationFromName(String) - 类 中的静态方法org.apache.spark.mllib.stat.correlation.Correlations
-
- getCount() - 类 中的方法org.apache.spark.storage.CountingWritableChannel
-
- getCurrentProcessingTimeMs() - 接口 中的方法org.apache.spark.sql.streaming.GroupState
-
Get the current processing time as milliseconds in epoch time.
- getCurrentUserGroups(SparkConf, String) - 类 中的静态方法org.apache.spark.util.Utils
-
- getCurrentUserName() - 类 中的静态方法org.apache.spark.util.Utils
-
Returns the current user name.
- getCurrentWatermarkMs() - 接口 中的方法org.apache.spark.sql.streaming.GroupState
-
Get the current event time watermark as milliseconds in epoch time.
- getData(Row) - 类 中的静态方法org.apache.spark.ml.image.ImageSchema
-
Gets the image data
- getDatabase(String) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
Get the database with the specified name.
- getDate(int) - 接口 中的方法org.apache.spark.sql.Row
-
Returns the value at position i of date type as java.sql.Date.
- getDecimal(int) - 接口 中的方法org.apache.spark.sql.Row
-
Returns the value at position i of decimal type as java.math.BigDecimal.
- getDecimal(int, int, int) - 类 中的方法org.apache.spark.sql.vectorized.ArrowColumnVector
-
- getDecimal(int, int, int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarArray
-
- getDecimal(int, int, int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarRow
-
- getDecimal(int, int, int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnVector
-
Returns the decimal type value for rowId.
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.OneVsRest
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.OneVsRestModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeans
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixture
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.clustering.KMeans
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.clustering.KMeansModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.evaluation.ClusteringEvaluator
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.evaluation.RegressionEvaluator
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.Binarizer
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.Bucketizer
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelector
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.ColumnPruner
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizer
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizerModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.DCT
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.ElementwiseProduct
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.FeatureHasher
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.HashingTF
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.IDF
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.IDFModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.Imputer
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.ImputerModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.IndexToString
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.Interaction
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.MaxAbsScaler
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.MaxAbsScalerModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSH
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSHModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScaler
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScalerModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.NGram
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.Normalizer
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoder
-
已过时。
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderEstimator
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.PCA
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.PCAModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.PolynomialExpansion
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.QuantileDiscretizer
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.RegexTokenizer
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.RFormula
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.SQLTransformer
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.StandardScaler
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.StandardScalerModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.StopWordsRemover
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.StringIndexer
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.StringIndexerModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.Tokenizer
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.VectorAssembler
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.VectorAttributeRewriter
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexer
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexerModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.VectorSizeHint
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.VectorSlicer
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.Word2Vec
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.Word2VecModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowth
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowthModel
-
- getDefault(Param<T>) - 接口 中的方法org.apache.spark.ml.param.Params
-
Gets the default value of a parameter.
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.Pipeline
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.PipelineModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.recommendation.ALSModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegression
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegressionModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidator
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidatorModel
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplit
-
- getDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplitModel
-
- getDefaultPropertiesFile(Map<String, String>) - 类 中的静态方法org.apache.spark.util.Utils
-
Return the path of the default Spark properties file.
- getDefaultSession() - 类 中的静态方法org.apache.spark.sql.SparkSession
-
Returns the default SparkSession that is returned by the builder.
- getDegree() - 类 中的方法org.apache.spark.ml.feature.PolynomialExpansion
-
- getDenseSizeInBytes() - 接口 中的方法org.apache.spark.ml.linalg.Matrix
-
Gets the size of the dense representation of this `Matrix`.
- getDependencies() - 类 中的方法org.apache.spark.rdd.CoGroupedRDD
-
- getDependencies() - 类 中的方法org.apache.spark.rdd.ShuffledRDD
-
- getDependencies() - 类 中的方法org.apache.spark.rdd.UnionRDD
-
- getDeprecatedConfig(String, Map<String, String>) - 类 中的静态方法org.apache.spark.SparkConf
-
Looks for available deprecated keys for the given config option, and return the first
value available.
- getDocConcentration() - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- getDocConcentration() - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- getDocConcentration() - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- getDocConcentration() - 类 中的方法org.apache.spark.mllib.clustering.LDA
-
Concentration parameter (commonly named "alpha") for the prior placed on documents'
distributions over topics ("theta").
- getDouble(String, double) - 类 中的方法org.apache.spark.SparkConf
-
Get a parameter as a double, falling back to a default if not set
- getDouble(int) - 接口 中的方法org.apache.spark.sql.Row
-
Returns the value at position i as a primitive double.
- getDouble(String, double) - 类 中的方法org.apache.spark.sql.sources.v2.DataSourceOptions
-
Returns the double value to which the specified key is mapped,
or defaultValue if there is no mapping for the key.
- getDouble(String) - 类 中的方法org.apache.spark.sql.types.Metadata
-
Gets a Double.
- getDouble(int) - 类 中的方法org.apache.spark.sql.vectorized.ArrowColumnVector
-
- getDouble(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarArray
-
- getDouble(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarRow
-
- getDouble(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnVector
-
Returns the double type value for rowId.
- getDoubleArray(String) - 类 中的方法org.apache.spark.sql.types.Metadata
-
Gets a Double array.
- getDoubles(int, int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnVector
-
Gets double type values from [rowId, rowId + count).
- getDropLast() - 类 中的方法org.apache.spark.ml.feature.OneHotEncoder
-
已过时。
- getDropLast() - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderEstimator
-
- getDropLast() - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderModel
-
- getDynamicAllocationInitialExecutors(SparkConf) - 类 中的静态方法org.apache.spark.util.Utils
-
Return the initial number of executors for dynamic allocation.
- getElasticNetParam() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- getElasticNetParam() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- getElasticNetParam() - 接口 中的方法org.apache.spark.ml.param.shared.HasElasticNetParam
-
- getElasticNetParam() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- getElasticNetParam() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- getEndOffset() - 接口 中的方法org.apache.spark.sql.sources.v2.reader.streaming.MicroBatchReader
-
Return the specified (if explicitly set through setOffsetRange) or inferred end offset
for this reader.
- getEndTimeEpoch() - 类 中的方法org.apache.spark.status.api.v1.ApplicationAttemptInfo
-
- getEpsilon() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- getEpsilon() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- getEpsilon() - 类 中的方法org.apache.spark.mllib.clustering.KMeans
-
The distance threshold within which we've consider centers to have converged.
- getEstimator() - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidator
-
- getEstimator() - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidatorModel
-
- getEstimator() - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplit
-
- getEstimator() - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplitModel
-
- getEstimatorParamMaps() - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidator
-
- getEstimatorParamMaps() - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidatorModel
-
- getEstimatorParamMaps() - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplit
-
- getEstimatorParamMaps() - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplitModel
-
- getEvaluator() - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidator
-
- getEvaluator() - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidatorModel
-
- getEvaluator() - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplit
-
- getEvaluator() - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplitModel
-
- GetExecutorEndpointRef(String) - 类 的构造器org.apache.spark.storage.BlockManagerMessages.GetExecutorEndpointRef
-
- GetExecutorEndpointRef$() - 类 的构造器org.apache.spark.storage.BlockManagerMessages.GetExecutorEndpointRef$
-
- getExecutorEnv() - 类 中的方法org.apache.spark.SparkConf
-
Get all executor environment variables set on this SparkConf
- getExecutorInfos() - 类 中的方法org.apache.spark.SparkStatusTracker
-
Returns information of all known executors, including host, port, cacheSize, numRunningTasks.
- GetExecutorLossReason(String) - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.GetExecutorLossReason
-
- GetExecutorLossReason$() - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.GetExecutorLossReason$
-
- getExecutorMemoryStatus() - 类 中的方法org.apache.spark.SparkContext
-
Return a map from the slave to the max memory available for caching and the remaining
memory available for caching.
- getExecutorStorageStatus() - 类 中的方法org.apache.spark.SparkContext
-
- getFamily() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- getFamily() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- getFamily() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- getFamily() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- getFdr() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelector
-
- getFdr() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- getFeatureIndex() - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegression
-
- getFeatureIndex() - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegressionModel
-
- getFeatureIndicesFromNames(StructField, String[]) - 类 中的静态方法org.apache.spark.ml.util.MetadataUtils
-
Takes a Vector column and a list of feature names, and returns the corresponding list of
feature indices in the column, in order.
- getFeaturesAndLabels(RFormulaModel, Dataset<?>) - 类 中的静态方法org.apache.spark.ml.r.RWrapperUtils
-
Get the feature names and original labels from the schema
of DataFrame transformed by RFormulaModel.
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.classification.OneVsRest
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.classification.OneVsRestModel
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeans
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixture
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.clustering.KMeans
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.clustering.KMeansModel
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.evaluation.ClusteringEvaluator
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelector
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.feature.RFormula
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.feature.RFormulaModel
-
- getFeaturesCol() - 接口 中的方法org.apache.spark.ml.param.shared.HasFeaturesCol
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegression
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegressionModel
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- getFeaturesCol() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- getFeatureSubsetStrategy() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- getFeatureSubsetStrategy() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- getFeatureSubsetStrategy() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- getFeatureSubsetStrategy() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- getFeatureSubsetStrategy() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- getFeatureSubsetStrategy() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- getFeatureSubsetStrategy() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- getFeatureSubsetStrategy() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- getField(String) - 类 中的方法org.apache.spark.sql.Column
-
An expression that gets a field by name in a StructType.
- getFileLength(File, SparkConf) - 类 中的静态方法org.apache.spark.util.Utils
-
Return the file length, if the file is compressed it returns the uncompressed file length.
- getFilePath(File, String) - 类 中的静态方法org.apache.spark.util.Utils
-
Return the absolute path of a file in the given directory.
- getFileReader(String, Option<Configuration>, boolean) - 类 中的静态方法org.apache.spark.sql.hive.orc.OrcFileOperator
-
Retrieves an ORC file reader from a given path.
- getFileSegmentLocations(String, long, long, Configuration) - 类 中的静态方法org.apache.spark.streaming.util.HdfsUtils
-
Get the locations of the HDFS blocks containing the given file segment.
- getFileSystemForPath(Path, Configuration) - 类 中的静态方法org.apache.spark.streaming.util.HdfsUtils
-
- getFinalStorageLevel() - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- getFinalValue() - 类 中的方法org.apache.spark.partial.PartialResult
-
Blocking method to wait for and return the final value.
- getFitIntercept() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- getFitIntercept() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- getFitIntercept() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- getFitIntercept() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- getFitIntercept() - 接口 中的方法org.apache.spark.ml.param.shared.HasFitIntercept
-
- getFitIntercept() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- getFitIntercept() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- getFitIntercept() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- getFitIntercept() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- getFitIntercept() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- getFitIntercept() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- getFloat(int) - 接口 中的方法org.apache.spark.sql.Row
-
Returns the value at position i as a primitive float.
- getFloat(int) - 类 中的方法org.apache.spark.sql.vectorized.ArrowColumnVector
-
- getFloat(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarArray
-
- getFloat(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarRow
-
- getFloat(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnVector
-
Returns the float type value for rowId.
- getFloats(int, int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnVector
-
Gets float type values from [rowId, rowId + count).
- getForceIndexLabel() - 类 中的静态方法org.apache.spark.ml.feature.RFormula
-
- getForceIndexLabel() - 类 中的静态方法org.apache.spark.ml.feature.RFormulaModel
-
- getFormattedClassName(Object) - 类 中的静态方法org.apache.spark.util.Utils
-
Return the class name of the given object, removing all dollar signs
- getFormula() - 类 中的静态方法org.apache.spark.ml.feature.RFormula
-
- getFormula() - 类 中的静态方法org.apache.spark.ml.feature.RFormulaModel
-
- getFpr() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelector
-
- getFpr() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- getFunction(String) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
Get the function with the specified name.
- getFunction(String, String) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
Get the function with the specified name.
- getFwe() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelector
-
- getFwe() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- getGaps() - 类 中的方法org.apache.spark.ml.feature.RegexTokenizer
-
- getGroups(String) - 接口 中的方法org.apache.spark.security.GroupMappingServiceProvider
-
Get the groups the user belongs to.
- getHadoopFileSystem(URI, Configuration) - 类 中的静态方法org.apache.spark.util.Utils
-
Return a Hadoop FileSystem with the scheme encoded in the given path.
- getHadoopFileSystem(String, Configuration) - 类 中的静态方法org.apache.spark.util.Utils
-
Return a Hadoop FileSystem with the scheme encoded in the given path.
- getHandleInvalid() - 类 中的静态方法org.apache.spark.ml.feature.Bucketizer
-
- getHandleInvalid() - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderEstimator
-
- getHandleInvalid() - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderModel
-
- getHandleInvalid() - 类 中的静态方法org.apache.spark.ml.feature.QuantileDiscretizer
-
- getHandleInvalid() - 类 中的静态方法org.apache.spark.ml.feature.RFormula
-
- getHandleInvalid() - 类 中的静态方法org.apache.spark.ml.feature.RFormulaModel
-
- getHandleInvalid() - 类 中的静态方法org.apache.spark.ml.feature.StringIndexer
-
- getHandleInvalid() - 类 中的静态方法org.apache.spark.ml.feature.StringIndexerModel
-
- getHandleInvalid() - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexer
-
- getHandleInvalid() - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexerModel
-
- getHandleInvalid() - 类 中的静态方法org.apache.spark.ml.feature.VectorSizeHint
-
- getHandleInvalid() - 接口 中的方法org.apache.spark.ml.param.shared.HasHandleInvalid
-
- getHeight(Row) - 类 中的静态方法org.apache.spark.ml.image.ImageSchema
-
Gets the height of the image
- getHiveWriteCompression(TableDesc, SQLConf) - 类 中的静态方法org.apache.spark.sql.hive.execution.HiveOptions
-
- getImplicitPrefs() - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- getImpurity() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- getImpurity() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- getImpurity() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- getImpurity() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- getImpurity() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- getImpurity() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- getImpurity() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- getImpurity() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- getImpurity() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- getImpurity() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- getImpurity() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- getImpurity() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- getImpurity() - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- getIndexValue(String, Object) - 类 中的方法org.apache.spark.util.kvstore.KVTypeInfo
-
- getIndices() - 类 中的方法org.apache.spark.ml.feature.VectorSlicer
-
- getInitializationMode() - 类 中的方法org.apache.spark.mllib.clustering.KMeans
-
The initialization algorithm.
- getInitializationSteps() - 类 中的方法org.apache.spark.mllib.clustering.KMeans
-
Number of steps for the k-means|| initialization mode
- getInitialModel() - 类 中的方法org.apache.spark.mllib.clustering.GaussianMixture
-
Return the user supplied initial GMM, if supplied
- getInitialPositionInStream(int) - 类 中的方法org.apache.spark.streaming.kinesis.KinesisUtilsPythonHelper
-
- getInitialTargetExecutorNumber(SparkConf, int) - 类 中的静态方法org.apache.spark.scheduler.cluster.SchedulerBackendUtils
-
Getting the initial target number of executors depends on whether dynamic allocation is
enabled.
- getInitialWeights() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- getInitMode() - 类 中的静态方法org.apache.spark.ml.clustering.KMeans
-
- getInitMode() - 类 中的静态方法org.apache.spark.ml.clustering.KMeansModel
-
- getInitSteps() - 类 中的静态方法org.apache.spark.ml.clustering.KMeans
-
- getInitSteps() - 类 中的静态方法org.apache.spark.ml.clustering.KMeansModel
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.Binarizer
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.Bucketizer
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizer
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizerModel
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.DCT
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.ElementwiseProduct
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.HashingTF
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.IDF
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.IDFModel
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.IndexToString
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.MaxAbsScaler
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.MaxAbsScalerModel
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSH
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSHModel
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScaler
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScalerModel
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.NGram
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.Normalizer
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoder
-
已过时。
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.PCA
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.PCAModel
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.PolynomialExpansion
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.QuantileDiscretizer
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.RegexTokenizer
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.StandardScaler
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.StandardScalerModel
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.StopWordsRemover
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.StringIndexer
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.StringIndexerModel
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.Tokenizer
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexer
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexerModel
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.VectorSizeHint
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.VectorSlicer
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.Word2Vec
-
- getInputCol() - 类 中的静态方法org.apache.spark.ml.feature.Word2VecModel
-
- getInputCol() - 接口 中的方法org.apache.spark.ml.param.shared.HasInputCol
-
- getInputCols() - 类 中的静态方法org.apache.spark.ml.feature.Bucketizer
-
- getInputCols() - 类 中的静态方法org.apache.spark.ml.feature.FeatureHasher
-
- getInputCols() - 类 中的静态方法org.apache.spark.ml.feature.Imputer
-
- getInputCols() - 类 中的静态方法org.apache.spark.ml.feature.ImputerModel
-
- getInputCols() - 类 中的静态方法org.apache.spark.ml.feature.Interaction
-
- getInputCols() - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderEstimator
-
- getInputCols() - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderModel
-
- getInputCols() - 类 中的静态方法org.apache.spark.ml.feature.QuantileDiscretizer
-
- getInputCols() - 类 中的静态方法org.apache.spark.ml.feature.VectorAssembler
-
- getInputCols() - 接口 中的方法org.apache.spark.ml.param.shared.HasInputCols
-
- getInputFilePath() - 类 中的静态方法org.apache.spark.rdd.InputFileBlockHolder
-
Returns the holding file name or empty string if it is unknown.
- getInputStream(String, Configuration) - 类 中的静态方法org.apache.spark.streaming.util.HdfsUtils
-
- getInt(String, int) - 类 中的方法org.apache.spark.SparkConf
-
Get a parameter as an integer, falling back to a default if not set
- getInt(int) - 接口 中的方法org.apache.spark.sql.Row
-
Returns the value at position i as a primitive int.
- getInt(String, int) - 类 中的方法org.apache.spark.sql.sources.v2.DataSourceOptions
-
Returns the integer value to which the specified key is mapped,
or defaultValue if there is no mapping for the key.
- getInt(int) - 类 中的方法org.apache.spark.sql.vectorized.ArrowColumnVector
-
- getInt(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarArray
-
- getInt(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarRow
-
- getInt(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnVector
-
Returns the int type value for rowId.
- getIntermediateStorageLevel() - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- getInterval(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarArray
-
- getInterval(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarRow
-
- getInterval(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnVector
-
Returns the calendar interval type value for rowId.
- getInts(int, int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnVector
-
Gets int type values from [rowId, rowId + count).
- getInverse() - 类 中的方法org.apache.spark.ml.feature.DCT
-
- getIsotonic() - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegression
-
- getIsotonic() - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegressionModel
-
- getItem(Object) - 类 中的方法org.apache.spark.sql.Column
-
An expression that gets an item at position ordinal out of an array,
or gets a value by key key in a MapType.
- getItemCol() - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- getItemCol() - 类 中的静态方法org.apache.spark.ml.recommendation.ALSModel
-
- getItemsCol() - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowth
-
- getItemsCol() - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowthModel
-
- getIteratorSize(Iterator<T>) - 类 中的静态方法org.apache.spark.util.Utils
-
Counts the number of elements of an iterator using a while loop rather than calling
TraversableOnce.size() because it uses a for loop, which is slightly slower
in the current version of Scala.
- getIteratorZipWithIndex(Iterator<T>, long) - 类 中的静态方法org.apache.spark.util.Utils
-
Generate a zipWithIndex iterator, avoid index value overflowing problem
in scala's zipWithIndex
- getJavaMap(int) - 接口 中的方法org.apache.spark.sql.Row
-
Returns the value at position i of array type as a java.util.Map.
- getJavaSparkContext(SparkSession) - 类 中的静态方法org.apache.spark.sql.api.r.SQLUtils
-
- getJDBCType(DataType) - 类 中的方法org.apache.spark.sql.jdbc.AggregatedDialect
-
- getJDBCType(DataType) - 类 中的静态方法org.apache.spark.sql.jdbc.DB2Dialect
-
- getJDBCType(DataType) - 类 中的静态方法org.apache.spark.sql.jdbc.DerbyDialect
-
- getJDBCType(DataType) - 类 中的方法org.apache.spark.sql.jdbc.JdbcDialect
-
Retrieve the jdbc / sql type for a given datatype.
- getJDBCType(DataType) - 类 中的静态方法org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- getJDBCType(DataType) - 类 中的静态方法org.apache.spark.sql.jdbc.MySQLDialect
-
- getJDBCType(DataType) - 类 中的静态方法org.apache.spark.sql.jdbc.NoopDialect
-
- getJDBCType(DataType) - 类 中的静态方法org.apache.spark.sql.jdbc.OracleDialect
-
- getJDBCType(DataType) - 类 中的静态方法org.apache.spark.sql.jdbc.PostgresDialect
-
- getJDBCType(DataType) - 类 中的静态方法org.apache.spark.sql.jdbc.TeradataDialect
-
- getJobIdsForGroup(String) - 类 中的方法org.apache.spark.api.java.JavaSparkStatusTracker
-
Return a list of all known jobs in a particular job group.
- getJobIdsForGroup(String) - 类 中的方法org.apache.spark.SparkStatusTracker
-
Return a list of all known jobs in a particular job group.
- getJobInfo(int) - 类 中的方法org.apache.spark.api.java.JavaSparkStatusTracker
-
Returns job information, or null if the job info could not be found or was garbage collected.
- getJobInfo(int) - 类 中的方法org.apache.spark.SparkStatusTracker
-
Returns job information, or None if the job info could not be found or was garbage collected.
- getK() - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeans
-
- getK() - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
- getK() - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- getK() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixture
-
- getK() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- getK() - 类 中的静态方法org.apache.spark.ml.clustering.KMeans
-
- getK() - 类 中的静态方法org.apache.spark.ml.clustering.KMeansModel
-
- getK() - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- getK() - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- getK() - 类 中的静态方法org.apache.spark.ml.feature.PCA
-
- getK() - 类 中的静态方法org.apache.spark.ml.feature.PCAModel
-
- getK() - 类 中的方法org.apache.spark.mllib.clustering.BisectingKMeans
-
Gets the desired number of leaf clusters.
- getK() - 类 中的方法org.apache.spark.mllib.clustering.GaussianMixture
-
Return the number of Gaussians in the mixture model
- getK() - 类 中的方法org.apache.spark.mllib.clustering.KMeans
-
Number of clusters to create (k).
- getK() - 类 中的方法org.apache.spark.mllib.clustering.LDA
-
Number of topics to infer, i.e., the number of soft cluster centers.
- getKappa() - 类 中的方法org.apache.spark.mllib.clustering.OnlineLDAOptimizer
-
Learning rate: exponential decay rate
- getKeepLastCheckpoint() - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- getKeepLastCheckpoint() - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- getKeepLastCheckpoint() - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- getKeepLastCheckpoint() - 类 中的方法org.apache.spark.mllib.clustering.EMLDAOptimizer
-
If using checkpointing, this indicates whether to keep the last checkpoint (vs clean up).
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.classification.OneVsRest
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.classification.OneVsRestModel
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.evaluation.RegressionEvaluator
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelector
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.feature.RFormula
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.feature.RFormulaModel
-
- getLabelCol() - 接口 中的方法org.apache.spark.ml.param.shared.HasLabelCol
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegression
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegressionModel
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- getLabelCol() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- getLabels() - 类 中的方法org.apache.spark.ml.feature.IndexToString
-
- getLambda() - 类 中的方法org.apache.spark.mllib.classification.NaiveBayes
-
Get the smoothing parameter.
- getLastUpdatedEpoch() - 类 中的方法org.apache.spark.status.api.v1.ApplicationAttemptInfo
-
- getLayers() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- getLDAModel(double[]) - 接口 中的方法org.apache.spark.mllib.clustering.LDAOptimizer
-
- getLearningDecay() - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- getLearningDecay() - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- getLearningDecay() - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- getLearningOffset() - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- getLearningOffset() - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- getLearningOffset() - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- getLearningRate() - 类 中的方法org.apache.spark.mllib.tree.configuration.BoostingStrategy
-
- getLeastGroupHash(String) - 类 中的方法org.apache.spark.rdd.DefaultPartitionCoalescer
-
Sorts and gets the least element of the list associated with key in groupHash
The returned PartitionGroup is the least loaded of all groups that represent the machine "key"
- getLength() - 类 中的静态方法org.apache.spark.rdd.InputFileBlockHolder
-
Returns the length of the block being read, or -1 if it is unknown.
- getLink() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- getLink() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- getLinkPower() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- getLinkPower() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- getLinkPredictionCol() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- getLinkPredictionCol() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- getList(int) - 接口 中的方法org.apache.spark.sql.Row
-
Returns the value at position i of array type as java.util.List.
- getLocalDir(SparkConf) - 类 中的静态方法org.apache.spark.util.Utils
-
Get the path of a temporary directory.
- getLocalizedMessage() - 异常错误 中的静态方法org.apache.spark.sql.AnalysisException
-
- getLocalProperty(String) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
Get a local property set in this thread, or null if it is missing.
- getLocalProperty(String) - 类 中的方法org.apache.spark.SparkContext
-
Get a local property set in this thread, or null if it is missing.
- getLocalProperty(String) - 类 中的方法org.apache.spark.TaskContext
-
Get a local property set upstream in the driver, or null if it is missing.
- getLocalUserJarsForShell(SparkConf) - 类 中的静态方法org.apache.spark.util.Utils
-
Return the local jar files which will be added to REPL's classpath.
- GetLocations(BlockId) - 类 的构造器org.apache.spark.storage.BlockManagerMessages.GetLocations
-
- GetLocations$() - 类 的构造器org.apache.spark.storage.BlockManagerMessages.GetLocations$
-
- GetLocationsAndStatus(BlockId) - 类 的构造器org.apache.spark.storage.BlockManagerMessages.GetLocationsAndStatus
-
- GetLocationsAndStatus$() - 类 的构造器org.apache.spark.storage.BlockManagerMessages.GetLocationsAndStatus$
-
- GetLocationsMultipleBlockIds(BlockId[]) - 类 的构造器org.apache.spark.storage.BlockManagerMessages.GetLocationsMultipleBlockIds
-
- GetLocationsMultipleBlockIds$() - 类 的构造器org.apache.spark.storage.BlockManagerMessages.GetLocationsMultipleBlockIds$
-
- getLong(String, long) - 类 中的方法org.apache.spark.SparkConf
-
Get a parameter as a long, falling back to a default if not set
- getLong(int) - 接口 中的方法org.apache.spark.sql.Row
-
Returns the value at position i as a primitive long.
- getLong(String, long) - 类 中的方法org.apache.spark.sql.sources.v2.DataSourceOptions
-
Returns the long value to which the specified key is mapped,
or defaultValue if there is no mapping for the key.
- getLong(String) - 类 中的方法org.apache.spark.sql.types.Metadata
-
Gets a Long.
- getLong(int) - 类 中的方法org.apache.spark.sql.vectorized.ArrowColumnVector
-
- getLong(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarArray
-
- getLong(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarRow
-
- getLong(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnVector
-
Returns the long type value for rowId.
- getLongArray(String) - 类 中的方法org.apache.spark.sql.types.Metadata
-
Gets a Long array.
- getLongs(int, int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnVector
-
Gets long type values from [rowId, rowId + count).
- getLoss() - 接口 中的方法org.apache.spark.ml.param.shared.HasLoss
-
- getLoss() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- getLoss() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- getLoss() - 类 中的方法org.apache.spark.mllib.tree.configuration.BoostingStrategy
-
- getLossType() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- getLossType() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- getLossType() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- getLossType() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- getLowerBound(double, long, double) - 类 中的静态方法org.apache.spark.util.random.BinomialBounds
-
Returns a threshold p such that if we conduct n Bernoulli trials with success rate = p,
it is very unlikely to have more than fraction * n successes.
- getLowerBound(double) - 类 中的静态方法org.apache.spark.util.random.PoissonBounds
-
Returns a lambda such that Pr[X > s] is very small, where X ~ Pois(lambda).
- getLowerBoundsOnCoefficients() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- getLowerBoundsOnCoefficients() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- getLowerBoundsOnIntercepts() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- getLowerBoundsOnIntercepts() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- getMap(int) - 接口 中的方法org.apache.spark.sql.Row
-
Returns the value at position i of map type as a Scala Map.
- getMap(int) - 类 中的方法org.apache.spark.sql.vectorized.ArrowColumnVector
-
- getMap(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarArray
-
- getMap(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarRow
-
- getMap(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnVector
-
Returns the map type value for rowId.
- GetMatchingBlockIds(Function1<BlockId, Object>, boolean) - 类 的构造器org.apache.spark.storage.BlockManagerMessages.GetMatchingBlockIds
-
- GetMatchingBlockIds$() - 类 的构造器org.apache.spark.storage.BlockManagerMessages.GetMatchingBlockIds$
-
- getMax() - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScaler
-
- getMax() - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScalerModel
-
- getMaxBins() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- getMaxBins() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- getMaxBins() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- getMaxBins() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- getMaxBins() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- getMaxBins() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- getMaxBins() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- getMaxBins() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- getMaxBins() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- getMaxBins() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- getMaxBins() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- getMaxBins() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- getMaxBins() - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- getMaxCategories() - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexer
-
- getMaxCategories() - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexerModel
-
- getMaxDepth() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- getMaxDepth() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- getMaxDepth() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- getMaxDepth() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- getMaxDepth() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- getMaxDepth() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- getMaxDepth() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- getMaxDepth() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- getMaxDepth() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- getMaxDepth() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- getMaxDepth() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- getMaxDepth() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- getMaxDepth() - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- getMaxFailures(SparkConf, boolean) - 类 中的静态方法org.apache.spark.streaming.util.WriteAheadLogUtils
-
- getMaxIter() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- getMaxIter() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- getMaxIter() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- getMaxIter() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- getMaxIter() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- getMaxIter() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- getMaxIter() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- getMaxIter() - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeans
-
- getMaxIter() - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
- getMaxIter() - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- getMaxIter() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixture
-
- getMaxIter() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- getMaxIter() - 类 中的静态方法org.apache.spark.ml.clustering.KMeans
-
- getMaxIter() - 类 中的静态方法org.apache.spark.ml.clustering.KMeansModel
-
- getMaxIter() - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- getMaxIter() - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- getMaxIter() - 类 中的静态方法org.apache.spark.ml.feature.Word2Vec
-
- getMaxIter() - 类 中的静态方法org.apache.spark.ml.feature.Word2VecModel
-
- getMaxIter() - 接口 中的方法org.apache.spark.ml.param.shared.HasMaxIter
-
- getMaxIter() - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- getMaxIter() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- getMaxIter() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- getMaxIter() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- getMaxIter() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- getMaxIter() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- getMaxIter() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- getMaxIter() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- getMaxIter() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- getMaxIterations() - 类 中的方法org.apache.spark.mllib.clustering.BisectingKMeans
-
Gets the max number of k-means iterations to split clusters.
- getMaxIterations() - 类 中的方法org.apache.spark.mllib.clustering.GaussianMixture
-
Return the maximum number of iterations allowed
- getMaxIterations() - 类 中的方法org.apache.spark.mllib.clustering.KMeans
-
Maximum number of iterations allowed.
- getMaxIterations() - 类 中的方法org.apache.spark.mllib.clustering.LDA
-
Maximum number of iterations allowed.
- getMaxLocalProjDBSize() - 类 中的方法org.apache.spark.mllib.fpm.PrefixSpan
-
Gets the maximum number of items allowed in a projected database before local processing.
- getMaxMemoryInMB() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- getMaxMemoryInMB() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- getMaxMemoryInMB() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- getMaxMemoryInMB() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- getMaxMemoryInMB() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- getMaxMemoryInMB() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- getMaxMemoryInMB() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- getMaxMemoryInMB() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- getMaxMemoryInMB() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- getMaxMemoryInMB() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- getMaxMemoryInMB() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- getMaxMemoryInMB() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- getMaxMemoryInMB() - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- getMaxPatternLength() - 类 中的方法org.apache.spark.mllib.fpm.PrefixSpan
-
Gets the maximal pattern length (i.e. the length of the longest sequential pattern to consider.
- getMaxResultSize(SparkConf) - 类 中的静态方法org.apache.spark.util.Utils
-
- getMaxSentenceLength() - 类 中的静态方法org.apache.spark.ml.feature.Word2Vec
-
- getMaxSentenceLength() - 类 中的静态方法org.apache.spark.ml.feature.Word2VecModel
-
- GetMemoryStatus$() - 类 的构造器org.apache.spark.storage.BlockManagerMessages.GetMemoryStatus$
-
- getMessage() - 异常错误 中的方法org.apache.spark.sql.AnalysisException
-
- getMetadata(String) - 类 中的方法org.apache.spark.sql.types.Metadata
-
Gets a Metadata.
- getMetadata(Class<T>) - 类 中的方法org.apache.spark.util.kvstore.InMemoryStore
-
- getMetadata(Class<T>) - 接口 中的方法org.apache.spark.util.kvstore.KVStore
-
Returns app-specific metadata from the store, or null if it's not currently set.
- getMetadata(Class<T>) - 类 中的方法org.apache.spark.util.kvstore.LevelDB
-
- getMetadataArray(String) - 类 中的方法org.apache.spark.sql.types.Metadata
-
Gets a Metadata array.
- getMetricName() - 类 中的方法org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
-
- getMetricName() - 类 中的方法org.apache.spark.ml.evaluation.ClusteringEvaluator
-
- getMetricName() - 类 中的方法org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- getMetricName() - 类 中的方法org.apache.spark.ml.evaluation.RegressionEvaluator
-
- getMetricsSources(String) - 类 中的方法org.apache.spark.TaskContext
-
::DeveloperApi::
Returns all metrics sources with the given name which are associated with the instance
which runs the task.
- getMin() - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScaler
-
- getMin() - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScalerModel
-
- getMinConfidence() - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowth
-
- getMinConfidence() - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowthModel
-
- getMinCount() - 类 中的静态方法org.apache.spark.ml.feature.Word2Vec
-
- getMinCount() - 类 中的静态方法org.apache.spark.ml.feature.Word2VecModel
-
- getMinDF() - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizer
-
- getMinDF() - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizerModel
-
- getMinDivisibleClusterSize() - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeans
-
- getMinDivisibleClusterSize() - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
- getMinDivisibleClusterSize() - 类 中的方法org.apache.spark.mllib.clustering.BisectingKMeans
-
Gets the minimum number of points (if greater than or equal to 1.0) or the minimum proportion
of points (if less than 1.0) of a divisible cluster.
- getMinDocFreq() - 类 中的静态方法org.apache.spark.ml.feature.IDF
-
- getMinDocFreq() - 类 中的静态方法org.apache.spark.ml.feature.IDFModel
-
- getMiniBatchFraction() - 类 中的方法org.apache.spark.mllib.clustering.OnlineLDAOptimizer
-
Mini-batch fraction, which sets the fraction of document sampled and used in each iteration
- getMinInfoGain() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- getMinInfoGain() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- getMinInfoGain() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- getMinInfoGain() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- getMinInfoGain() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- getMinInfoGain() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- getMinInfoGain() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- getMinInfoGain() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- getMinInfoGain() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- getMinInfoGain() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- getMinInfoGain() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- getMinInfoGain() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- getMinInfoGain() - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- getMinInstancesPerNode() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- getMinInstancesPerNode() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- getMinInstancesPerNode() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- getMinInstancesPerNode() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- getMinInstancesPerNode() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- getMinInstancesPerNode() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- getMinInstancesPerNode() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- getMinInstancesPerNode() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- getMinInstancesPerNode() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- getMinInstancesPerNode() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- getMinInstancesPerNode() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- getMinInstancesPerNode() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- getMinInstancesPerNode() - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- getMinSupport() - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowth
-
- getMinSupport() - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowthModel
-
- getMinSupport() - 类 中的方法org.apache.spark.mllib.fpm.PrefixSpan
-
Get the minimal support (i.e. the frequency of occurrence before a pattern is considered
frequent).
- getMinTF() - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizer
-
- getMinTF() - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizerModel
-
- getMinTokenLength() - 类 中的方法org.apache.spark.ml.feature.RegexTokenizer
-
- getMissingValue() - 类 中的静态方法org.apache.spark.ml.feature.Imputer
-
- getMissingValue() - 类 中的静态方法org.apache.spark.ml.feature.ImputerModel
-
- getMode(Row) - 类 中的静态方法org.apache.spark.ml.image.ImageSchema
-
Gets the OpenCV representation as an int
- getModelType() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- getModelType() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- getModelType() - 类 中的方法org.apache.spark.mllib.classification.NaiveBayes
-
Get the model type.
- getN() - 类 中的方法org.apache.spark.ml.feature.NGram
-
- getNames() - 类 中的方法org.apache.spark.ml.feature.VectorSlicer
-
- getNChannels(Row) - 类 中的静态方法org.apache.spark.ml.image.ImageSchema
-
Gets the number of channels in the image
- getNode(int, Node) - 类 中的静态方法org.apache.spark.mllib.tree.model.Node
-
Traces down from a root node to get the node with the given node index.
- getNonnegative() - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- getNumBuckets() - 类 中的静态方法org.apache.spark.ml.feature.QuantileDiscretizer
-
- getNumBucketsArray() - 类 中的静态方法org.apache.spark.ml.feature.QuantileDiscretizer
-
- getNumClasses(StructField) - 类 中的静态方法org.apache.spark.ml.util.MetadataUtils
-
Examine a schema to identify the number of classes in a label column.
- getNumClasses() - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- getNumFeatures() - 类 中的方法org.apache.spark.ml.feature.FeatureHasher
-
- getNumFeatures() - 类 中的方法org.apache.spark.ml.feature.HashingTF
-
- getNumFeatures() - 类 中的静态方法org.apache.spark.mllib.classification.LogisticRegressionWithSGD
-
- getNumFeatures() - 类 中的静态方法org.apache.spark.mllib.classification.SVMWithSGD
-
- getNumFeatures() - 类 中的方法org.apache.spark.mllib.regression.GeneralizedLinearAlgorithm
-
The dimension of training features.
- getNumFeatures() - 类 中的静态方法org.apache.spark.mllib.regression.LassoWithSGD
-
- getNumFeatures() - 类 中的静态方法org.apache.spark.mllib.regression.LinearRegressionWithSGD
-
- getNumFeatures() - 类 中的静态方法org.apache.spark.mllib.regression.RidgeRegressionWithSGD
-
- getNumFolds() - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidator
-
- getNumFolds() - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidatorModel
-
- getNumHashTables() - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- getNumHashTables() - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
-
- getNumHashTables() - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSH
-
- getNumHashTables() - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSHModel
-
- getNumItemBlocks() - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- getNumIterations() - 类 中的方法org.apache.spark.mllib.tree.configuration.BoostingStrategy
-
- getNumObjFields() - 类 中的方法org.apache.spark.serializer.SerializationDebugger.ObjectStreamClassMethods
-
- getNumPartitions() - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- getNumPartitions() - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- getNumPartitions() - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- getNumPartitions() - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Return the number of partitions in this RDD.
- getNumPartitions() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- getNumPartitions() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- getNumPartitions() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- getNumPartitions() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- getNumPartitions() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- getNumPartitions() - 类 中的静态方法org.apache.spark.ml.feature.Word2Vec
-
- getNumPartitions() - 类 中的静态方法org.apache.spark.ml.feature.Word2VecModel
-
- getNumPartitions() - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowth
-
- getNumPartitions() - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowthModel
-
- getNumPartitions() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- getNumPartitions() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- getNumPartitions() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- getNumPartitions() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- getNumPartitions() - 类 中的方法org.apache.spark.rdd.RDD
-
Returns the number of partitions of this RDD.
- getNumPartitions() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- getNumTopFeatures() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelector
-
- getNumTopFeatures() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- getNumTrees() - 类 中的方法org.apache.spark.ml.classification.GBTClassificationModel
-
Number of trees in ensemble
- getNumTrees() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- getNumTrees() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- getNumTrees() - 类 中的方法org.apache.spark.ml.regression.GBTRegressionModel
-
Number of trees in ensemble
- getNumTrees() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- getNumTrees() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- getNumUserBlocks() - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- getNumValues() - 类 中的方法org.apache.spark.ml.attribute.NominalAttribute
-
Get the number of values, either from numValues or from values.
- getObjectInspector(String, Option<Configuration>) - 类 中的静态方法org.apache.spark.sql.hive.orc.OrcFileOperator
-
- getObjFieldValues(Object, Object[]) - 类 中的方法org.apache.spark.serializer.SerializationDebugger.ObjectStreamClassMethods
-
- getOffset() - 接口 中的方法org.apache.spark.sql.sources.v2.reader.streaming.ContinuousDataReader
-
Get the offset of the current record, or the start offset if no records have been read.
- getOffsetCol() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- getOffsetCol() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- getOptimizeDocConcentration() - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- getOptimizeDocConcentration() - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- getOptimizeDocConcentration() - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- getOptimizeDocConcentration() - 类 中的方法org.apache.spark.mllib.clustering.OnlineLDAOptimizer
-
Optimize docConcentration, indicates whether docConcentration (Dirichlet parameter for
document-topic distribution) will be optimized during training.
- getOptimizer() - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- getOptimizer() - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- getOptimizer() - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- getOptimizer() - 类 中的方法org.apache.spark.mllib.clustering.LDA
-
:: DeveloperApi ::
LDAOptimizer used to perform the actual calculation
- getOption(String) - 类 中的方法org.apache.spark.SparkConf
-
Get a parameter as an Option
- getOption(String) - 类 中的方法org.apache.spark.sql.RuntimeConfig
-
Returns the value of Spark runtime configuration property for the given key.
- getOption() - 接口 中的方法org.apache.spark.sql.streaming.GroupState
-
Get the state value as a scala Option.
- getOption() - 类 中的方法org.apache.spark.streaming.State
-
Get the state as a scala.Option.
- getOrCreate(SparkConf) - 类 中的静态方法org.apache.spark.SparkContext
-
This function may be used to get or instantiate a SparkContext and register it as a
singleton object.
- getOrCreate() - 类 中的静态方法org.apache.spark.SparkContext
-
This function may be used to get or instantiate a SparkContext and register it as a
singleton object.
- getOrCreate() - 类 中的方法org.apache.spark.sql.SparkSession.Builder
-
Gets an existing
SparkSession or, if there is no existing one, creates a new
one based on the options set in this builder.
- getOrCreate(SparkContext) - 类 中的静态方法org.apache.spark.sql.SQLContext
-
- getOrCreate(String, Function0<JavaStreamingContext>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaStreamingContext
-
Either recreate a StreamingContext from checkpoint data or create a new StreamingContext.
- getOrCreate(String, Function0<JavaStreamingContext>, Configuration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaStreamingContext
-
Either recreate a StreamingContext from checkpoint data or create a new StreamingContext.
- getOrCreate(String, Function0<JavaStreamingContext>, Configuration, boolean) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaStreamingContext
-
Either recreate a StreamingContext from checkpoint data or create a new StreamingContext.
- getOrCreate(String, Function0<StreamingContext>, Configuration, boolean) - 类 中的静态方法org.apache.spark.streaming.StreamingContext
-
Either recreate a StreamingContext from checkpoint data or create a new StreamingContext.
- getOrCreateSparkSession(JavaSparkContext, Map<Object, Object>, boolean) - 类 中的静态方法org.apache.spark.sql.api.r.SQLUtils
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.OneVsRest
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.OneVsRestModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeans
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixture
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.clustering.KMeans
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.clustering.KMeansModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.evaluation.ClusteringEvaluator
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.evaluation.RegressionEvaluator
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.Binarizer
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.Bucketizer
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelector
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.ColumnPruner
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizer
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizerModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.DCT
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.ElementwiseProduct
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.FeatureHasher
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.HashingTF
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.IDF
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.IDFModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.Imputer
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.ImputerModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.IndexToString
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.Interaction
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.MaxAbsScaler
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.MaxAbsScalerModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSH
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSHModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScaler
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScalerModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.NGram
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.Normalizer
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoder
-
已过时。
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderEstimator
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.PCA
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.PCAModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.PolynomialExpansion
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.QuantileDiscretizer
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.RegexTokenizer
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.RFormula
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.SQLTransformer
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.StandardScaler
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.StandardScalerModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.StopWordsRemover
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.StringIndexer
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.StringIndexerModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.Tokenizer
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.VectorAssembler
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.VectorAttributeRewriter
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexer
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexerModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.VectorSizeHint
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.VectorSlicer
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.Word2Vec
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.feature.Word2VecModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowth
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowthModel
-
- getOrDefault(Param<T>) - 接口 中的方法org.apache.spark.ml.param.Params
-
Gets the value of a param in the embedded param map or its default value.
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.Pipeline
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.PipelineModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.recommendation.ALSModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegression
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegressionModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidator
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidatorModel
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplit
-
- getOrDefault(Param<T>) - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplitModel
-
- getOrElse(Param<T>, T) - 类 中的方法org.apache.spark.ml.param.ParamMap
-
Returns the value associated with a param or a default value.
- getOrigin(Row) - 类 中的静态方法org.apache.spark.ml.image.ImageSchema
-
Gets the origin of the image
- getOutputAttrGroupFromData(Dataset<?>, Seq<String>, Seq<String>, boolean) - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderCommon
-
This method is called when we want to generate AttributeGroup from actual data for
one-hot encoder.
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.Binarizer
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.Bucketizer
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelector
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizer
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizerModel
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.DCT
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.ElementwiseProduct
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.FeatureHasher
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.HashingTF
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.IDF
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.IDFModel
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.IndexToString
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.Interaction
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.MaxAbsScaler
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.MaxAbsScalerModel
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSH
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSHModel
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScaler
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScalerModel
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.NGram
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.Normalizer
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoder
-
已过时。
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.PCA
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.PCAModel
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.PolynomialExpansion
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.QuantileDiscretizer
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.RegexTokenizer
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.StandardScaler
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.StandardScalerModel
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.StopWordsRemover
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.StringIndexer
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.StringIndexerModel
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.Tokenizer
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.VectorAssembler
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexer
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexerModel
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.VectorSlicer
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.Word2Vec
-
- getOutputCol() - 类 中的静态方法org.apache.spark.ml.feature.Word2VecModel
-
- getOutputCol() - 接口 中的方法org.apache.spark.ml.param.shared.HasOutputCol
-
- getOutputCols() - 类 中的静态方法org.apache.spark.ml.feature.Bucketizer
-
- getOutputCols() - 类 中的静态方法org.apache.spark.ml.feature.Imputer
-
- getOutputCols() - 类 中的静态方法org.apache.spark.ml.feature.ImputerModel
-
- getOutputCols() - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderEstimator
-
- getOutputCols() - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderModel
-
- getOutputCols() - 类 中的静态方法org.apache.spark.ml.feature.QuantileDiscretizer
-
- getOutputCols() - 接口 中的方法org.apache.spark.ml.param.shared.HasOutputCols
-
- getOutputStream(String, Configuration) - 类 中的静态方法org.apache.spark.streaming.util.HdfsUtils
-
- getP() - 类 中的方法org.apache.spark.ml.feature.Normalizer
-
- getParallelism() - 类 中的静态方法org.apache.spark.ml.classification.OneVsRest
-
- getParallelism() - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidator
-
- getParallelism() - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplit
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.classification.OneVsRest
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.classification.OneVsRestModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeans
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixture
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.clustering.KMeans
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.clustering.KMeansModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.evaluation.ClusteringEvaluator
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.evaluation.RegressionEvaluator
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.Binarizer
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.Bucketizer
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelector
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.ColumnPruner
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizer
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizerModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.DCT
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.ElementwiseProduct
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.FeatureHasher
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.HashingTF
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.IDF
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.IDFModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.Imputer
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.ImputerModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.IndexToString
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.Interaction
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.MaxAbsScaler
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.MaxAbsScalerModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSH
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSHModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScaler
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScalerModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.NGram
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.Normalizer
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoder
-
已过时。
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderEstimator
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.PCA
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.PCAModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.PolynomialExpansion
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.QuantileDiscretizer
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.RegexTokenizer
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.RFormula
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.SQLTransformer
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.StandardScaler
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.StandardScalerModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.StopWordsRemover
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.StringIndexer
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.StringIndexerModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.Tokenizer
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.VectorAssembler
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.VectorAttributeRewriter
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexer
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexerModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.VectorSizeHint
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.VectorSlicer
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.Word2Vec
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.feature.Word2VecModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowth
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowthModel
-
- getParam(String) - 接口 中的方法org.apache.spark.ml.param.Params
-
Gets a param by its name.
- getParam(String) - 类 中的静态方法org.apache.spark.ml.Pipeline
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.PipelineModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.recommendation.ALSModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegression
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegressionModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidator
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidatorModel
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplit
-
- getParam(String) - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplitModel
-
- getParents(int) - 类 中的方法org.apache.spark.NarrowDependency
-
Get the parent partitions for a child partition.
- getParents(int) - 类 中的方法org.apache.spark.OneToOneDependency
-
- getParents(int) - 类 中的方法org.apache.spark.RangeDependency
-
- getPartition(long, long, int) - 类 中的方法org.apache.spark.graphx.PartitionStrategy.CanonicalRandomVertexCut$
-
- getPartition(long, long, int) - 类 中的方法org.apache.spark.graphx.PartitionStrategy.EdgePartition1D$
-
- getPartition(long, long, int) - 类 中的方法org.apache.spark.graphx.PartitionStrategy.EdgePartition2D$
-
- getPartition(long, long, int) - 接口 中的方法org.apache.spark.graphx.PartitionStrategy
-
Returns the partition number for a given edge.
- getPartition(long, long, int) - 类 中的方法org.apache.spark.graphx.PartitionStrategy.RandomVertexCut$
-
- getPartition(Object) - 类 中的方法org.apache.spark.HashPartitioner
-
- getPartition(Object) - 类 中的方法org.apache.spark.Partitioner
-
- getPartition(Object) - 类 中的方法org.apache.spark.RangePartitioner
-
- getPartitionId() - 类 中的静态方法org.apache.spark.TaskContext
-
Returns the partition id of currently active TaskContext.
- getPartitions() - 类 中的方法org.apache.spark.api.r.BaseRRDD
-
- getPartitions() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- getPartitions() - 类 中的方法org.apache.spark.rdd.CoGroupedRDD
-
- getPartitions() - 类 中的方法org.apache.spark.rdd.DefaultPartitionCoalescer
-
- getPartitions() - 类 中的方法org.apache.spark.rdd.HadoopRDD
-
- getPartitions() - 类 中的方法org.apache.spark.rdd.JdbcRDD
-
- getPartitions() - 类 中的方法org.apache.spark.rdd.NewHadoopRDD
-
- getPartitions() - 类 中的方法org.apache.spark.rdd.ShuffledRDD
-
- getPartitions() - 类 中的方法org.apache.spark.rdd.UnionRDD
-
- getPath() - 类 中的方法org.apache.spark.input.PortableDataStream
-
- getPattern() - 类 中的方法org.apache.spark.ml.feature.RegexTokenizer
-
- GetPeers(BlockManagerId) - 类 的构造器org.apache.spark.storage.BlockManagerMessages.GetPeers
-
- GetPeers$() - 类 的构造器org.apache.spark.storage.BlockManagerMessages.GetPeers$
-
- getPercentile() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelector
-
- getPercentile() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- getPersistentRDDs() - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
Returns a Java map of JavaRDDs that have marked themselves as persistent via cache() call.
- getPersistentRDDs() - 类 中的方法org.apache.spark.SparkContext
-
Returns an immutable map of RDDs that have marked themselves as persistent via cache() call.
- getPoissonSamplingFunction(RDD<Tuple2<K, V>>, Map<K, Object>, boolean, long, ClassTag<K>, ClassTag<V>) - 类 中的静态方法org.apache.spark.util.random.StratifiedSamplingUtils
-
Return the per partition sampling function used for sampling with replacement.
- getPoolForName(String) - 类 中的方法org.apache.spark.SparkContext
-
:: DeveloperApi ::
Return the pool associated with the given name, if one exists
- getPosition() - 类 中的方法org.apache.spark.streaming.kinesis.KinesisInitialPositions.AtTimestamp
-
- getPosition() - 类 中的方法org.apache.spark.streaming.kinesis.KinesisInitialPositions.Latest
-
- getPosition() - 类 中的方法org.apache.spark.streaming.kinesis.KinesisInitialPositions.TrimHorizon
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.OneVsRest
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.OneVsRestModel
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeans
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixture
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.clustering.KMeans
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.clustering.KMeansModel
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.evaluation.ClusteringEvaluator
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.evaluation.RegressionEvaluator
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowth
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowthModel
-
- getPredictionCol() - 接口 中的方法org.apache.spark.ml.param.shared.HasPredictionCol
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.recommendation.ALSModel
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegression
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegressionModel
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- getPredictionCol() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- getPreferredLocations(Partition) - 类 中的方法org.apache.spark.rdd.HadoopRDD
-
- getPreferredLocations(Partition) - 类 中的方法org.apache.spark.rdd.NewHadoopRDD
-
- getPreferredLocations(Partition) - 类 中的方法org.apache.spark.rdd.UnionRDD
-
- getProbabilityCol() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- getProbabilityCol() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- getProbabilityCol() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- getProbabilityCol() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- getProbabilityCol() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- getProbabilityCol() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- getProbabilityCol() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- getProbabilityCol() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- getProbabilityCol() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- getProbabilityCol() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- getProbabilityCol() - 类 中的静态方法org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- getProbabilityCol() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- getProbabilityCol() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- getProbabilityCol() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixture
-
- getProbabilityCol() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- getProbabilityCol() - 接口 中的方法org.apache.spark.ml.param.shared.HasProbabilityCol
-
- getProcessName() - 类 中的静态方法org.apache.spark.util.Utils
-
Returns the name of this JVM process.
- getPropertiesFromFile(String) - 类 中的静态方法org.apache.spark.util.Utils
-
Load properties present in the given file.
- getQuantileCalculationStrategy() - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- getQuantileProbabilities() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- getQuantileProbabilities() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- getQuantilesCol() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- getQuantilesCol() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- getRandomSample(Seq<T>, int, Random) - 类 中的静态方法org.apache.spark.storage.BlockReplicationUtils
-
Get a random sample of size m from the elems
- getRank() - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- getRatingCol() - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- getRawPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- getRawPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- getRawPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- getRawPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- getRawPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- getRawPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- getRawPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- getRawPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- getRawPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- getRawPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- getRawPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- getRawPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- getRawPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- getRawPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- getRawPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- getRawPredictionCol() - 类 中的静态方法org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
-
- getRawPredictionCol() - 接口 中的方法org.apache.spark.ml.param.shared.HasRawPredictionCol
-
- getRddBlockLocations(int, Seq<StorageStatus>) - 类 中的静态方法org.apache.spark.storage.StorageUtils
-
Return a mapping from block ID to its locations for each block that belongs to the given RDD.
- getRDDStorageInfo() - 类 中的方法org.apache.spark.SparkContext
-
:: DeveloperApi ::
Return information about what RDDs are cached, if they are in mem or on disk, how much space
they take, etc.
- getReceiver() - 类 中的方法org.apache.spark.streaming.dstream.ReceiverInputDStream
-
Gets the receiver object that will be sent to the worker nodes
to receive data.
- getRegParam() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- getRegParam() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- getRegParam() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- getRegParam() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- getRegParam() - 接口 中的方法org.apache.spark.ml.param.shared.HasRegParam
-
- getRegParam() - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- getRegParam() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- getRegParam() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- getRegParam() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- getRegParam() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- getRelativeError() - 类 中的静态方法org.apache.spark.ml.feature.QuantileDiscretizer
-
- getRollingIntervalSecs(SparkConf, boolean) - 类 中的静态方法org.apache.spark.streaming.util.WriteAheadLogUtils
-
- getRootDirectory() - 类 中的静态方法org.apache.spark.SparkFiles
-
Get the root directory that contains files added through SparkContext.addFile().
- getRow(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarBatch
-
Returns the row in this batch at `rowId`.
- getRuns() - 类 中的方法org.apache.spark.mllib.clustering.KMeans
-
- getScalingVec() - 类 中的方法org.apache.spark.ml.feature.ElementwiseProduct
-
- getSchedulingMode() - 类 中的方法org.apache.spark.SparkContext
-
Return current scheduling mode
- getSchemaQuery(String) - 类 中的方法org.apache.spark.sql.jdbc.AggregatedDialect
-
- getSchemaQuery(String) - 类 中的静态方法org.apache.spark.sql.jdbc.DB2Dialect
-
- getSchemaQuery(String) - 类 中的静态方法org.apache.spark.sql.jdbc.DerbyDialect
-
- getSchemaQuery(String) - 类 中的方法org.apache.spark.sql.jdbc.JdbcDialect
-
The SQL query that should be used to discover the schema of a table.
- getSchemaQuery(String) - 类 中的静态方法org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- getSchemaQuery(String) - 类 中的静态方法org.apache.spark.sql.jdbc.MySQLDialect
-
- getSchemaQuery(String) - 类 中的静态方法org.apache.spark.sql.jdbc.NoopDialect
-
- getSchemaQuery(String) - 类 中的静态方法org.apache.spark.sql.jdbc.OracleDialect
-
- getSchemaQuery(String) - 类 中的静态方法org.apache.spark.sql.jdbc.PostgresDialect
-
- getSchemaQuery(String) - 类 中的静态方法org.apache.spark.sql.jdbc.TeradataDialect
-
- getSeed() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- getSeed() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- getSeed() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- getSeed() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- getSeed() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- getSeed() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- getSeed() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- getSeed() - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeans
-
- getSeed() - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
- getSeed() - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- getSeed() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixture
-
- getSeed() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- getSeed() - 类 中的静态方法org.apache.spark.ml.clustering.KMeans
-
- getSeed() - 类 中的静态方法org.apache.spark.ml.clustering.KMeansModel
-
- getSeed() - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- getSeed() - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- getSeed() - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- getSeed() - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSH
-
- getSeed() - 类 中的静态方法org.apache.spark.ml.feature.Word2Vec
-
- getSeed() - 类 中的静态方法org.apache.spark.ml.feature.Word2VecModel
-
- getSeed() - 接口 中的方法org.apache.spark.ml.param.shared.HasSeed
-
- getSeed() - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- getSeed() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- getSeed() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- getSeed() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- getSeed() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- getSeed() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- getSeed() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- getSeed() - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidator
-
- getSeed() - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidatorModel
-
- getSeed() - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplit
-
- getSeed() - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplitModel
-
- getSeed() - 类 中的方法org.apache.spark.mllib.clustering.BisectingKMeans
-
Gets the random seed.
- getSeed() - 类 中的方法org.apache.spark.mllib.clustering.GaussianMixture
-
Return the random seed
- getSeed() - 类 中的方法org.apache.spark.mllib.clustering.KMeans
-
The random seed for cluster initialization.
- getSeed() - 类 中的方法org.apache.spark.mllib.clustering.LDA
-
Random seed for cluster initialization.
- getSelectorType() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelector
-
- getSelectorType() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- getSeq(int) - 接口 中的方法org.apache.spark.sql.Row
-
Returns the value at position i of array type as a Scala Seq.
- getSeqOp(boolean, Map<K, Object>, StratifiedSamplingUtils.RandomDataGenerator, Option<Map<K, Object>>) - 类 中的静态方法org.apache.spark.util.random.StratifiedSamplingUtils
-
Returns the function used by aggregate to collect sampling statistics for each partition.
- getSessionConf(SparkSession) - 类 中的静态方法org.apache.spark.sql.api.r.SQLUtils
-
- getShort(int) - 接口 中的方法org.apache.spark.sql.Row
-
Returns the value at position i as a primitive short.
- getShort(int) - 类 中的方法org.apache.spark.sql.vectorized.ArrowColumnVector
-
- getShort(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarArray
-
- getShort(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarRow
-
- getShort(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnVector
-
Returns the short type value for rowId.
- getShorts(int, int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnVector
-
Gets short type values from [rowId, rowId + count).
- getSimpleMessage() - 异常错误 中的方法org.apache.spark.sql.AnalysisException
-
- getSize() - 类 中的方法org.apache.spark.ml.feature.VectorSizeHint
-
group getParam
- getSizeAsBytes(String) - 类 中的方法org.apache.spark.SparkConf
-
Get a size parameter as bytes; throws a NoSuchElementException if it's not set.
- getSizeAsBytes(String, String) - 类 中的方法org.apache.spark.SparkConf
-
Get a size parameter as bytes, falling back to a default if not set.
- getSizeAsBytes(String, long) - 类 中的方法org.apache.spark.SparkConf
-
Get a size parameter as bytes, falling back to a default if not set.
- getSizeAsGb(String) - 类 中的方法org.apache.spark.SparkConf
-
Get a size parameter as Gibibytes; throws a NoSuchElementException if it's not set.
- getSizeAsGb(String, String) - 类 中的方法org.apache.spark.SparkConf
-
Get a size parameter as Gibibytes, falling back to a default if not set.
- getSizeAsKb(String) - 类 中的方法org.apache.spark.SparkConf
-
Get a size parameter as Kibibytes; throws a NoSuchElementException if it's not set.
- getSizeAsKb(String, String) - 类 中的方法org.apache.spark.SparkConf
-
Get a size parameter as Kibibytes, falling back to a default if not set.
- getSizeAsMb(String) - 类 中的方法org.apache.spark.SparkConf
-
Get a size parameter as Mebibytes; throws a NoSuchElementException if it's not set.
- getSizeAsMb(String, String) - 类 中的方法org.apache.spark.SparkConf
-
Get a size parameter as Mebibytes, falling back to a default if not set.
- getSizeInBytes() - 接口 中的方法org.apache.spark.ml.linalg.Matrix
-
Gets the current size in bytes of this `Matrix`.
- getSlotDescs() - 类 中的方法org.apache.spark.serializer.SerializationDebugger.ObjectStreamClassMethods
-
- getSmoothing() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- getSmoothing() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- getSolver() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- getSolver() - 接口 中的方法org.apache.spark.ml.param.shared.HasSolver
-
- getSolver() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- getSolver() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- getSolver() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- getSolver() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- getSparkClassLoader() - 类 中的静态方法org.apache.spark.util.Utils
-
Get the ClassLoader which loaded Spark.
- getSparkHome() - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
Get Spark's home location from either a value set through the constructor,
or the spark.home Java property, or the SPARK_HOME environment variable
(in that order of preference).
- getSparkOrYarnConfig(SparkConf, String, String) - 类 中的静态方法org.apache.spark.util.Utils
-
Return the value of a config either through the SparkConf or the Hadoop configuration
if this is Yarn mode.
- getSparseSizeInBytes(boolean) - 接口 中的方法org.apache.spark.ml.linalg.Matrix
-
Gets the size of the minimal sparse representation of this `Matrix`.
- getSplit() - 类 中的方法org.apache.spark.ml.tree.DecisionTreeModelReadWrite.SplitData
-
- getSplits() - 类 中的方法org.apache.spark.ml.feature.Bucketizer
-
- getSplitsArray() - 类 中的方法org.apache.spark.ml.feature.Bucketizer
-
- getStackTrace() - 异常错误 中的静态方法org.apache.spark.sql.AnalysisException
-
- getStageInfo(int) - 类 中的方法org.apache.spark.api.java.JavaSparkStatusTracker
-
Returns stage information, or null if the stage info could not be found or was
garbage collected.
- getStageInfo(int) - 类 中的方法org.apache.spark.SparkStatusTracker
-
Returns stage information, or None if the stage info could not be found or was
garbage collected.
- getStagePath(String, int, int, String) - 类 中的方法org.apache.spark.ml.Pipeline.SharedReadWrite$
-
Get path for saving the given stage.
- getStages() - 类 中的方法org.apache.spark.ml.Pipeline
-
- getStandardization() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- getStandardization() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- getStandardization() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- getStandardization() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- getStandardization() - 接口 中的方法org.apache.spark.ml.param.shared.HasStandardization
-
- getStandardization() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- getStandardization() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- getStartOffset() - 类 中的静态方法org.apache.spark.rdd.InputFileBlockHolder
-
Returns the starting offset of the block currently being read, or -1 if it is unknown.
- getStartOffset() - 接口 中的方法org.apache.spark.sql.sources.v2.reader.streaming.ContinuousReader
-
Return the specified or inferred start offset for this reader.
- getStartOffset() - 接口 中的方法org.apache.spark.sql.sources.v2.reader.streaming.MicroBatchReader
-
Returns the specified (if explicitly set through setOffsetRange) or inferred start offset
for this reader.
- getStartTimeEpoch() - 类 中的方法org.apache.spark.status.api.v1.ApplicationAttemptInfo
-
- getState() - 接口 中的方法org.apache.spark.launcher.SparkAppHandle
-
Returns the current application state.
- getState() - 类 中的方法org.apache.spark.streaming.api.java.JavaStreamingContext
-
:: DeveloperApi ::
Return the current state of the context.
- getState() - 类 中的方法org.apache.spark.streaming.StreamingContext
-
:: DeveloperApi ::
Return the current state of the context.
- getStatement() - 类 中的方法org.apache.spark.ml.feature.SQLTransformer
-
- getStatistics() - 接口 中的方法org.apache.spark.sql.sources.v2.reader.SupportsReportStatistics
-
Returns the basic statistics of this data source.
- getStderr(Process, long) - 类 中的静态方法org.apache.spark.util.Utils
-
Return the stderr of a process after waiting for the process to terminate.
- getStepSize() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- getStepSize() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- getStepSize() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- getStepSize() - 类 中的静态方法org.apache.spark.ml.feature.Word2Vec
-
- getStepSize() - 类 中的静态方法org.apache.spark.ml.feature.Word2VecModel
-
- getStepSize() - 接口 中的方法org.apache.spark.ml.param.shared.HasStepSize
-
- getStepSize() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- getStepSize() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- getStopWords() - 类 中的方法org.apache.spark.ml.feature.StopWordsRemover
-
- getStorageLevel() - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- getStorageLevel() - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- getStorageLevel() - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- getStorageLevel() - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Get the RDD's current storage level, or StorageLevel.NONE if none is set.
- getStorageLevel() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- getStorageLevel() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- getStorageLevel() - 类 中的方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- getStorageLevel() - 类 中的方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- getStorageLevel() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- getStorageLevel() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- getStorageLevel() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- getStorageLevel() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- getStorageLevel() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- getStorageLevel() - 类 中的方法org.apache.spark.rdd.RDD
-
Get the RDD's current storage level, or StorageLevel.NONE if none is set.
- getStorageLevel() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- GetStorageStatus$() - 类 的构造器org.apache.spark.storage.BlockManagerMessages.GetStorageStatus$
-
- getStrategy() - 类 中的静态方法org.apache.spark.ml.feature.Imputer
-
- getStrategy() - 类 中的静态方法org.apache.spark.ml.feature.ImputerModel
-
- getString(int) - 接口 中的方法org.apache.spark.sql.Row
-
Returns the value at position i as a String object.
- getString(String) - 类 中的方法org.apache.spark.sql.types.Metadata
-
Gets a String.
- getStringArray(String) - 类 中的方法org.apache.spark.sql.types.Metadata
-
Gets a String array.
- getStringIndexerOrderType() - 类 中的静态方法org.apache.spark.ml.feature.RFormula
-
- getStringIndexerOrderType() - 类 中的静态方法org.apache.spark.ml.feature.RFormulaModel
-
- getStringOrderType() - 类 中的静态方法org.apache.spark.ml.feature.StringIndexer
-
- getStringOrderType() - 类 中的静态方法org.apache.spark.ml.feature.StringIndexerModel
-
- getStruct(int) - 接口 中的方法org.apache.spark.sql.Row
-
Returns the value at position i of struct type as a
Row object.
- getStruct(int, int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarArray
-
- getStruct(int, int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarRow
-
- getStruct(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnVector
-
Returns the struct type value for rowId.
- getSubsamplingRate() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- getSubsamplingRate() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- getSubsamplingRate() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- getSubsamplingRate() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- getSubsamplingRate() - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- getSubsamplingRate() - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- getSubsamplingRate() - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- getSubsamplingRate() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- getSubsamplingRate() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- getSubsamplingRate() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- getSubsamplingRate() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- getSubsamplingRate() - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- getSuppressed() - 异常错误 中的静态方法org.apache.spark.sql.AnalysisException
-
- getSystemProperties() - 类 中的静态方法org.apache.spark.util.Utils
-
Returns the system properties map that is thread-safe to iterator over.
- getTable(String) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
Get the table or view with the specified name.
- getTable(String, String) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
Get the table or view with the specified name in the specified database.
- getTableExistsQuery(String) - 类 中的方法org.apache.spark.sql.jdbc.AggregatedDialect
-
- getTableExistsQuery(String) - 类 中的静态方法org.apache.spark.sql.jdbc.DB2Dialect
-
- getTableExistsQuery(String) - 类 中的静态方法org.apache.spark.sql.jdbc.DerbyDialect
-
- getTableExistsQuery(String) - 类 中的方法org.apache.spark.sql.jdbc.JdbcDialect
-
Get the SQL query that should be used to find if the given table exists.
- getTableExistsQuery(String) - 类 中的静态方法org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- getTableExistsQuery(String) - 类 中的静态方法org.apache.spark.sql.jdbc.MySQLDialect
-
- getTableExistsQuery(String) - 类 中的静态方法org.apache.spark.sql.jdbc.NoopDialect
-
- getTableExistsQuery(String) - 类 中的静态方法org.apache.spark.sql.jdbc.OracleDialect
-
- getTableExistsQuery(String) - 类 中的静态方法org.apache.spark.sql.jdbc.PostgresDialect
-
- getTableExistsQuery(String) - 类 中的静态方法org.apache.spark.sql.jdbc.TeradataDialect
-
- getTableNames(SparkSession, String) - 类 中的静态方法org.apache.spark.sql.api.r.SQLUtils
-
- getTables(SparkSession, String) - 类 中的静态方法org.apache.spark.sql.api.r.SQLUtils
-
- getTau0() - 类 中的方法org.apache.spark.mllib.clustering.OnlineLDAOptimizer
-
A (positive) learning parameter that downweights early iterations.
- getThreadDump() - 类 中的静态方法org.apache.spark.util.Utils
-
Return a thread dump of all threads' stacktraces.
- getThreadDumpForThread(long) - 类 中的静态方法org.apache.spark.util.Utils
-
- getThreshold() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- getThreshold() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- getThreshold() - 类 中的方法org.apache.spark.ml.classification.LogisticRegression
-
- getThreshold() - 类 中的方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- getThreshold() - 类 中的方法org.apache.spark.ml.feature.Binarizer
-
- getThreshold() - 接口 中的方法org.apache.spark.ml.param.shared.HasThreshold
-
- getThreshold() - 类 中的方法org.apache.spark.mllib.classification.LogisticRegressionModel
-
Returns the threshold (if any) used for converting raw prediction scores into 0/1 predictions.
- getThreshold() - 类 中的方法org.apache.spark.mllib.classification.SVMModel
-
Returns the threshold (if any) used for converting raw prediction scores into 0/1 predictions.
- getThresholds() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- getThresholds() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- getThresholds() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- getThresholds() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- getThresholds() - 类 中的方法org.apache.spark.ml.classification.LogisticRegression
-
- getThresholds() - 类 中的方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- getThresholds() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- getThresholds() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- getThresholds() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- getThresholds() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- getThresholds() - 类 中的静态方法org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- getThresholds() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- getThresholds() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- getThresholds() - 接口 中的方法org.apache.spark.ml.param.shared.HasThresholds
-
- getTimeAsMs(String) - 类 中的方法org.apache.spark.SparkConf
-
Get a time parameter as milliseconds; throws a NoSuchElementException if it's not set.
- getTimeAsMs(String, String) - 类 中的方法org.apache.spark.SparkConf
-
Get a time parameter as milliseconds, falling back to a default if not set.
- getTimeAsSeconds(String) - 类 中的方法org.apache.spark.SparkConf
-
Get a time parameter as seconds; throws a NoSuchElementException if it's not set.
- getTimeAsSeconds(String, String) - 类 中的方法org.apache.spark.SparkConf
-
Get a time parameter as seconds, falling back to a default if not set.
- getTimestamp(int) - 接口 中的方法org.apache.spark.sql.Row
-
Returns the value at position i of date type as java.sql.Timestamp.
- getTimestamp() - 类 中的方法org.apache.spark.streaming.kinesis.KinesisInitialPositions.AtTimestamp
-
- getTimeZoneOffset() - 类 中的静态方法org.apache.spark.ui.UIUtils
-
- GETTING_RESULT_TIME() - 类 中的静态方法org.apache.spark.status.TaskIndexNames
-
- GETTING_RESULT_TIME() - 类 中的静态方法org.apache.spark.ui.jobs.TaskDetailsClassNames
-
- GETTING_RESULT_TIME() - 类 中的静态方法org.apache.spark.ui.ToolTips
-
- gettingResult() - 类 中的方法org.apache.spark.scheduler.TaskInfo
-
- gettingResultTime() - 类 中的方法org.apache.spark.scheduler.TaskInfo
-
The time when the task started remotely getting the result.
- gettingResultTime() - 类 中的方法org.apache.spark.status.api.v1.TaskMetricDistributions
-
- gettingResultTime(TaskData) - 类 中的静态方法org.apache.spark.status.AppStatusUtils
-
- gettingResultTime(long, long, long) - 类 中的静态方法org.apache.spark.status.AppStatusUtils
-
- getTol() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- getTol() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- getTol() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- getTol() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- getTol() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- getTol() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixture
-
- getTol() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- getTol() - 类 中的静态方法org.apache.spark.ml.clustering.KMeans
-
- getTol() - 类 中的静态方法org.apache.spark.ml.clustering.KMeansModel
-
- getTol() - 接口 中的方法org.apache.spark.ml.param.shared.HasTol
-
- getTol() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- getTol() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- getTol() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- getTol() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- getTol() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- getTol() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- getToLowercase() - 类 中的方法org.apache.spark.ml.feature.RegexTokenizer
-
- getTopicConcentration() - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- getTopicConcentration() - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- getTopicConcentration() - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- getTopicConcentration() - 类 中的方法org.apache.spark.mllib.clustering.LDA
-
Concentration parameter (commonly named "beta" or "eta") for the prior placed on topics'
distributions over terms.
- getTopicDistributionCol() - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- getTopicDistributionCol() - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- getTopicDistributionCol() - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- getTopologyForHost(String) - 类 中的方法org.apache.spark.storage.DefaultTopologyMapper
-
- getTopologyForHost(String) - 类 中的方法org.apache.spark.storage.FileBasedTopologyMapper
-
- getTopologyForHost(String) - 类 中的方法org.apache.spark.storage.TopologyMapper
-
Gets the topology information given the host name
- getTrainRatio() - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplit
-
- getTrainRatio() - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplitModel
-
- getTreeStrategy() - 类 中的方法org.apache.spark.mllib.tree.configuration.BoostingStrategy
-
- getTruncateQuery(String) - 类 中的方法org.apache.spark.sql.jdbc.AggregatedDialect
-
- getTruncateQuery(String) - 类 中的静态方法org.apache.spark.sql.jdbc.DB2Dialect
-
- getTruncateQuery(String) - 类 中的静态方法org.apache.spark.sql.jdbc.DerbyDialect
-
- getTruncateQuery(String) - 类 中的方法org.apache.spark.sql.jdbc.JdbcDialect
-
The SQL query that should be used to truncate a table.
- getTruncateQuery(String) - 类 中的静态方法org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- getTruncateQuery(String) - 类 中的静态方法org.apache.spark.sql.jdbc.MySQLDialect
-
- getTruncateQuery(String) - 类 中的静态方法org.apache.spark.sql.jdbc.NoopDialect
-
- getTruncateQuery(String) - 类 中的静态方法org.apache.spark.sql.jdbc.OracleDialect
-
- getTruncateQuery(String) - 类 中的静态方法org.apache.spark.sql.jdbc.PostgresDialect
-
The SQL query used to truncate a table.
- getTruncateQuery(String) - 类 中的静态方法org.apache.spark.sql.jdbc.TeradataDialect
-
- getUDTFor(String) - 类 中的静态方法org.apache.spark.sql.types.UDTRegistration
-
Returns the Class of UserDefinedType for the name of a given user class.
- getUidMap(Params) - 类 中的静态方法org.apache.spark.ml.util.MetaAlgorithmReadWrite
-
Examine the given estimator (which may be a compound estimator) and extract a mapping
from UIDs to corresponding Params instances.
- getUiRoot(ServletContext) - 类 中的静态方法org.apache.spark.status.api.v1.UIRootFromServletContext
-
- getUpperBound(double, long, double) - 类 中的静态方法org.apache.spark.util.random.BinomialBounds
-
Returns a threshold p such that if we conduct n Bernoulli trials with success rate = p,
it is very unlikely to have less than fraction * n successes.
- getUpperBound(double) - 类 中的静态方法org.apache.spark.util.random.PoissonBounds
-
Returns a lambda such that Pr[X < s] is very small, where X ~ Pois(lambda).
- getUpperBoundsOnCoefficients() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- getUpperBoundsOnCoefficients() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- getUpperBoundsOnIntercepts() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- getUpperBoundsOnIntercepts() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- getUsedTimeMs(long) - 类 中的静态方法org.apache.spark.util.Utils
-
Return the string to tell how long has passed in milliseconds.
- getUseNodeIdCache() - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- getUserCol() - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- getUserCol() - 类 中的静态方法org.apache.spark.ml.recommendation.ALSModel
-
- getUserJars(SparkConf) - 类 中的静态方法org.apache.spark.util.Utils
-
Return the jar files pointed by the "spark.jars" property.
- getUTF8String(int) - 类 中的方法org.apache.spark.sql.vectorized.ArrowColumnVector
-
- getUTF8String(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarArray
-
- getUTF8String(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarRow
-
- getUTF8String(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnVector
-
Returns the string type value for rowId.
- getValidationTol() - 类 中的方法org.apache.spark.mllib.tree.configuration.BoostingStrategy
-
- getValue(int) - 类 中的方法org.apache.spark.ml.attribute.NominalAttribute
-
Gets a value given its index.
- getValuesMap(Seq<String>) - 接口 中的方法org.apache.spark.sql.Row
-
Returns a Map consisting of names and values for the requested fieldNames
For primitive types if value is null it returns 'zero value' specific for primitive
ie. 0 for Int - use isNullAt to ensure that value is not null
- getVarianceCol() - 接口 中的方法org.apache.spark.ml.param.shared.HasVarianceCol
-
- getVarianceCol() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- getVarianceCol() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- getVariancePower() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- getVariancePower() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- getVectors() - 类 中的方法org.apache.spark.ml.feature.Word2VecModel
-
Returns a dataframe with two fields, "word" and "vector", with "word" being a String and
and the vector the DenseVector that it is mapped to.
- getVectors() - 类 中的方法org.apache.spark.mllib.feature.Word2VecModel
-
Returns a map of words to their vector representations.
- getVectorSize() - 类 中的静态方法org.apache.spark.ml.feature.Word2Vec
-
- getVectorSize() - 类 中的静态方法org.apache.spark.ml.feature.Word2VecModel
-
- getVocabSize() - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizer
-
- getVocabSize() - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizerModel
-
- getWeightCol() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- getWeightCol() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- getWeightCol() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- getWeightCol() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- getWeightCol() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- getWeightCol() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- getWeightCol() - 类 中的静态方法org.apache.spark.ml.classification.OneVsRest
-
- getWeightCol() - 类 中的静态方法org.apache.spark.ml.classification.OneVsRestModel
-
- getWeightCol() - 接口 中的方法org.apache.spark.ml.param.shared.HasWeightCol
-
- getWeightCol() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- getWeightCol() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- getWeightCol() - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegression
-
- getWeightCol() - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegressionModel
-
- getWeightCol() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- getWeightCol() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- getWidth(Row) - 类 中的静态方法org.apache.spark.ml.image.ImageSchema
-
Gets the width of the image
- getWindowSize() - 类 中的静态方法org.apache.spark.ml.feature.Word2Vec
-
- getWindowSize() - 类 中的静态方法org.apache.spark.ml.feature.Word2VecModel
-
- getWithMean() - 类 中的静态方法org.apache.spark.ml.feature.StandardScaler
-
- getWithMean() - 类 中的静态方法org.apache.spark.ml.feature.StandardScalerModel
-
- getWithStd() - 类 中的静态方法org.apache.spark.ml.feature.StandardScaler
-
- getWithStd() - 类 中的静态方法org.apache.spark.ml.feature.StandardScalerModel
-
- Gini - org.apache.spark.mllib.tree.impurity中的类
-
Class for calculating the Gini impurity
(http://en.wikipedia.org/wiki/Decision_tree_learning#Gini_impurity)
during multiclass classification.
- Gini() - 类 的构造器org.apache.spark.mllib.tree.impurity.Gini
-
- GLMClassificationModel - org.apache.spark.mllib.classification.impl中的类
-
Helper class for import/export of GLM classification models.
- GLMClassificationModel() - 类 的构造器org.apache.spark.mllib.classification.impl.GLMClassificationModel
-
- GLMClassificationModel.SaveLoadV1_0$ - org.apache.spark.mllib.classification.impl中的类
-
- GLMClassificationModel.SaveLoadV1_0$.Data - org.apache.spark.mllib.classification.impl中的类
-
Model data for import/export
- GLMRegressionModel - org.apache.spark.mllib.regression.impl中的类
-
Helper methods for import/export of GLM regression models.
- GLMRegressionModel() - 类 的构造器org.apache.spark.mllib.regression.impl.GLMRegressionModel
-
- GLMRegressionModel.SaveLoadV1_0$ - org.apache.spark.mllib.regression.impl中的类
-
- GLMRegressionModel.SaveLoadV1_0$.Data - org.apache.spark.mllib.regression.impl中的类
-
Model data for model import/export
- glom() - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- glom() - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- glom() - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- glom() - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Return an RDD created by coalescing all elements within each partition into an array.
- glom() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- glom() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- glom() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- glom() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- glom() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- glom() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- glom() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- glom() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- glom() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- glom() - 类 中的方法org.apache.spark.rdd.RDD
-
Return an RDD created by coalescing all elements within each partition into an array.
- glom() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- glom() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- glom() - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return a new DStream in which each RDD is generated by applying glom() to each RDD of
this DStream.
- glom() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- glom() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- glom() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- glom() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- glom() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- glom() - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Return a new DStream in which each RDD is generated by applying glom() to each RDD of
this DStream.
- goodnessOfFit() - 类 中的方法org.apache.spark.mllib.stat.test.ChiSqTest.NullHypothesis$
-
- grad() - 类 中的方法org.apache.spark.mllib.optimization.NNLS.Workspace
-
- gradient() - 类 中的方法org.apache.spark.ml.regression.AFTAggregator
-
- Gradient - org.apache.spark.mllib.optimization中的类
-
:: DeveloperApi ::
Class used to compute the gradient for a loss function, given a single data point.
- Gradient() - 类 的构造器org.apache.spark.mllib.optimization.Gradient
-
- gradient(double, double) - 类 中的静态方法org.apache.spark.mllib.tree.loss.AbsoluteError
-
Method to calculate the gradients for the gradient boosting calculation for least
absolute error calculation.
- gradient(double, double) - 类 中的静态方法org.apache.spark.mllib.tree.loss.LogLoss
-
Method to calculate the loss gradients for the gradient boosting calculation for binary
classification
The gradient with respect to F(x) is: - 4 y / (1 + exp(2 y F(x)))
- gradient(double, double) - 接口 中的方法org.apache.spark.mllib.tree.loss.Loss
-
Method to calculate the gradients for the gradient boosting calculation.
- gradient(double, double) - 类 中的静态方法org.apache.spark.mllib.tree.loss.SquaredError
-
Method to calculate the gradients for the gradient boosting calculation for least
squares error calculation.
- GradientBoostedTrees - org.apache.spark.ml.tree.impl中的类
-
- GradientBoostedTrees() - 类 的构造器org.apache.spark.ml.tree.impl.GradientBoostedTrees
-
- GradientBoostedTrees - org.apache.spark.mllib.tree中的类
-
- GradientBoostedTrees(BoostingStrategy) - 类 的构造器org.apache.spark.mllib.tree.GradientBoostedTrees
-
- GradientBoostedTreesModel - org.apache.spark.mllib.tree.model中的类
-
Represents a gradient boosted trees model.
- GradientBoostedTreesModel(Enumeration.Value, DecisionTreeModel[], double[]) - 类 的构造器org.apache.spark.mllib.tree.model.GradientBoostedTreesModel
-
- GradientDescent - org.apache.spark.mllib.optimization中的类
-
Class used to solve an optimization problem using Gradient Descent.
- Graph<VD,ED> - org.apache.spark.graphx中的类
-
The Graph abstractly represents a graph with arbitrary objects
associated with vertices and edges.
- GraphGenerators - org.apache.spark.graphx.util中的类
-
A collection of graph generating functions.
- GraphGenerators() - 类 的构造器org.apache.spark.graphx.util.GraphGenerators
-
- GraphImpl<VD,ED> - org.apache.spark.graphx.impl中的类
-
An implementation of
Graph to support computation on graphs.
- GraphLoader - org.apache.spark.graphx中的类
-
Provides utilities for loading
Graphs from files.
- GraphLoader() - 类 的构造器org.apache.spark.graphx.GraphLoader
-
- GraphOps<VD,ED> - org.apache.spark.graphx中的类
-
Contains additional functionality for
Graph.
- GraphOps(Graph<VD, ED>, ClassTag<VD>, ClassTag<ED>) - 类 的构造器org.apache.spark.graphx.GraphOps
-
- graphToGraphOps(Graph<VD, ED>, ClassTag<VD>, ClassTag<ED>) - 类 中的静态方法org.apache.spark.graphx.Graph
-
Implicitly extracts the
GraphOps member from a graph.
- GraphXUtils - org.apache.spark.graphx中的类
-
- GraphXUtils() - 类 的构造器org.apache.spark.graphx.GraphXUtils
-
- greater(Duration) - 类 中的方法org.apache.spark.streaming.Duration
-
- greater(Time) - 类 中的方法org.apache.spark.streaming.Time
-
- greaterEq(Duration) - 类 中的方法org.apache.spark.streaming.Duration
-
- greaterEq(Time) - 类 中的方法org.apache.spark.streaming.Time
-
- GreaterThan - org.apache.spark.sql.sources中的类
-
A filter that evaluates to true iff the attribute evaluates to a value
greater than value.
- GreaterThan(String, Object) - 类 的构造器org.apache.spark.sql.sources.GreaterThan
-
- GreaterThanOrEqual - org.apache.spark.sql.sources中的类
-
A filter that evaluates to true iff the attribute evaluates to a value
greater than or equal to value.
- GreaterThanOrEqual(String, Object) - 类 的构造器org.apache.spark.sql.sources.GreaterThanOrEqual
-
- greatest(Column...) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the greatest value of the list of values, skipping null values.
- greatest(String, String...) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the greatest value of the list of column names, skipping null values.
- greatest(Seq<Column>) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the greatest value of the list of values, skipping null values.
- greatest(String, Seq<String>) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the greatest value of the list of column names, skipping null values.
- gridGraph(SparkContext, int, int) - 类 中的静态方法org.apache.spark.graphx.util.GraphGenerators
-
Create rows by cols grid graph with each vertex connected to its
row+1 and col+1 neighbors.
- groupArr() - 类 中的方法org.apache.spark.rdd.DefaultPartitionCoalescer
-
- groupBy(Function<T, U>) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- groupBy(Function<T, U>, int) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- groupBy(Function<T, U>) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- groupBy(Function<T, U>, int) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- groupBy(Function<T, U>) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- groupBy(Function<T, U>, int) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- groupBy(Function<T, U>) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Return an RDD of grouped elements.
- groupBy(Function<T, U>, int) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Return an RDD of grouped elements.
- groupBy(Function1<T, K>, ClassTag<K>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- groupBy(Function1<T, K>, int, ClassTag<K>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- groupBy(Function1<T, K>, Partitioner, ClassTag<K>, Ordering<K>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- groupBy(Function1<T, K>, ClassTag<K>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- groupBy(Function1<T, K>, int, ClassTag<K>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- groupBy(Function1<T, K>, Partitioner, ClassTag<K>, Ordering<K>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- groupBy(Function1<T, K>, ClassTag<K>) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- groupBy(Function1<T, K>, int, ClassTag<K>) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- groupBy(Function1<T, K>, Partitioner, ClassTag<K>, Ordering<K>) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- groupBy(Function1<T, K>, ClassTag<K>) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- groupBy(Function1<T, K>, int, ClassTag<K>) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- groupBy(Function1<T, K>, Partitioner, ClassTag<K>, Ordering<K>) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- groupBy(Function1<T, K>, ClassTag<K>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- groupBy(Function1<T, K>, int, ClassTag<K>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- groupBy(Function1<T, K>, Partitioner, ClassTag<K>, Ordering<K>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- groupBy(Function1<T, K>, ClassTag<K>) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- groupBy(Function1<T, K>, int, ClassTag<K>) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- groupBy(Function1<T, K>, Partitioner, ClassTag<K>, Ordering<K>) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- groupBy(Function1<T, K>, ClassTag<K>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- groupBy(Function1<T, K>, int, ClassTag<K>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- groupBy(Function1<T, K>, Partitioner, ClassTag<K>, Ordering<K>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- groupBy(Function1<T, K>, ClassTag<K>) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- groupBy(Function1<T, K>, int, ClassTag<K>) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- groupBy(Function1<T, K>, Partitioner, ClassTag<K>, Ordering<K>) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- groupBy(Function1<T, K>, ClassTag<K>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- groupBy(Function1<T, K>, int, ClassTag<K>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- groupBy(Function1<T, K>, Partitioner, ClassTag<K>, Ordering<K>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- groupBy(Function1<T, K>, ClassTag<K>) - 类 中的方法org.apache.spark.rdd.RDD
-
Return an RDD of grouped items.
- groupBy(Function1<T, K>, int, ClassTag<K>) - 类 中的方法org.apache.spark.rdd.RDD
-
Return an RDD of grouped elements.
- groupBy(Function1<T, K>, Partitioner, ClassTag<K>, Ordering<K>) - 类 中的方法org.apache.spark.rdd.RDD
-
Return an RDD of grouped items.
- groupBy(Function1<T, K>, ClassTag<K>) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- groupBy(Function1<T, K>, int, ClassTag<K>) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- groupBy(Function1<T, K>, Partitioner, ClassTag<K>, Ordering<K>) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- groupBy(Column...) - 类 中的方法org.apache.spark.sql.Dataset
-
Groups the Dataset using the specified columns, so we can run aggregation on them.
- groupBy(String, String...) - 类 中的方法org.apache.spark.sql.Dataset
-
Groups the Dataset using the specified columns, so that we can run aggregation on them.
- groupBy(Seq<Column>) - 类 中的方法org.apache.spark.sql.Dataset
-
Groups the Dataset using the specified columns, so we can run aggregation on them.
- groupBy(String, Seq<String>) - 类 中的方法org.apache.spark.sql.Dataset
-
Groups the Dataset using the specified columns, so that we can run aggregation on them.
- groupBy(Function1<A, K>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- groupBy$default$4(Function1<T, K>, Partitioner) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- groupBy$default$4(Function1<T, K>, Partitioner) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- groupBy$default$4(Function1<T, K>, Partitioner) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- groupBy$default$4(Function1<T, K>, Partitioner) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- groupBy$default$4(Function1<T, K>, Partitioner) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- groupBy$default$4(Function1<T, K>, Partitioner) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- groupBy$default$4(Function1<T, K>, Partitioner) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- groupBy$default$4(Function1<T, K>, Partitioner) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- groupBy$default$4(Function1<T, K>, Partitioner) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- groupBy$default$4(Function1<T, K>, Partitioner) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- groupByKey(Partitioner) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Group the values for each key in the RDD into a single sequence.
- groupByKey(int) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Group the values for each key in the RDD into a single sequence.
- groupByKey() - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Group the values for each key in the RDD into a single sequence.
- groupByKey(Partitioner) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Group the values for each key in the RDD into a single sequence.
- groupByKey(int) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Group the values for each key in the RDD into a single sequence.
- groupByKey() - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Group the values for each key in the RDD into a single sequence.
- groupByKey(Function1<T, K>, Encoder<K>) - 类 中的方法org.apache.spark.sql.Dataset
-
:: Experimental ::
(Scala-specific)
Returns a
KeyValueGroupedDataset where the data is grouped by the given key
func.
- groupByKey(MapFunction<T, K>, Encoder<K>) - 类 中的方法org.apache.spark.sql.Dataset
-
:: Experimental ::
(Java-specific)
Returns a
KeyValueGroupedDataset where the data is grouped by the given key
func.
- groupByKey() - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying groupByKey to each RDD.
- groupByKey(int) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying groupByKey to each RDD.
- groupByKey(Partitioner) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying groupByKey on each RDD of this DStream.
- groupByKey() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- groupByKey(int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- groupByKey(Partitioner) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- groupByKey() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- groupByKey(int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- groupByKey(Partitioner) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- groupByKey() - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying groupByKey to each RDD.
- groupByKey(int) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying groupByKey to each RDD.
- groupByKey(Partitioner) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying groupByKey on each RDD.
- groupByKeyAndWindow(Duration) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying groupByKey over a sliding window.
- groupByKeyAndWindow(Duration, Duration) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying groupByKey over a sliding window.
- groupByKeyAndWindow(Duration, Duration, int) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying groupByKey over a sliding window on this DStream.
- groupByKeyAndWindow(Duration, Duration, Partitioner) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying groupByKey over a sliding window on this DStream.
- groupByKeyAndWindow(Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- groupByKeyAndWindow(Duration, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- groupByKeyAndWindow(Duration, Duration, int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- groupByKeyAndWindow(Duration, Duration, Partitioner) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- groupByKeyAndWindow(Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- groupByKeyAndWindow(Duration, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- groupByKeyAndWindow(Duration, Duration, int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- groupByKeyAndWindow(Duration, Duration, Partitioner) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- groupByKeyAndWindow(Duration) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying groupByKey over a sliding window.
- groupByKeyAndWindow(Duration, Duration) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying groupByKey over a sliding window.
- groupByKeyAndWindow(Duration, Duration, int) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying groupByKey over a sliding window on this DStream.
- groupByKeyAndWindow(Duration, Duration, Partitioner) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Create a new DStream by applying groupByKey over a sliding window on this DStream.
- GroupByType$() - 类 的构造器org.apache.spark.sql.RelationalGroupedDataset.GroupByType$
-
- grouped(int) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- groupEdges(Function2<ED, ED, ED>) - 类 中的方法org.apache.spark.graphx.Graph
-
Merges multiple edges between two vertices into a single edge.
- groupEdges(Function2<ED, ED, ED>) - 类 中的方法org.apache.spark.graphx.impl.GraphImpl
-
- groupHash() - 类 中的方法org.apache.spark.rdd.DefaultPartitionCoalescer
-
- grouping(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: indicates whether a specified column in a GROUP BY list is aggregated
or not, returns 1 for aggregated or 0 for not aggregated in the result set.
- grouping(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: indicates whether a specified column in a GROUP BY list is aggregated
or not, returns 1 for aggregated or 0 for not aggregated in the result set.
- grouping_id(Seq<Column>) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the level of grouping, equals to
(grouping(c1) <<; (n-1)) + (grouping(c2) <<; (n-2)) + ... + grouping(cn)
- grouping_id(String, Seq<String>) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the level of grouping, equals to
(grouping(c1) <<; (n-1)) + (grouping(c2) <<; (n-2)) + ... + grouping(cn)
- GroupMappingServiceProvider - org.apache.spark.security中的接口
-
This Spark trait is used for mapping a given userName to a set of groups which it belongs to.
- GroupState<S> - org.apache.spark.sql.streaming中的接口
-
:: Experimental ::
Wrapper class for interacting with per-group state data in mapGroupsWithState and
flatMapGroupsWithState operations on KeyValueGroupedDataset.
- GroupStateTimeout - org.apache.spark.sql.streaming中的类
-
Represents the type of timeouts possible for the Dataset operations
`mapGroupsWithState` and `flatMapGroupsWithState`.
- GroupStateTimeout() - 类 的构造器org.apache.spark.sql.streaming.GroupStateTimeout
-
- groupWith(JavaPairRDD<K, W>) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Alias for cogroup.
- groupWith(JavaPairRDD<K, W1>, JavaPairRDD<K, W2>) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Alias for cogroup.
- groupWith(JavaPairRDD<K, W1>, JavaPairRDD<K, W2>, JavaPairRDD<K, W3>) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Alias for cogroup.
- groupWith(RDD<Tuple2<K, W>>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Alias for cogroup.
- groupWith(RDD<Tuple2<K, W1>>, RDD<Tuple2<K, W2>>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Alias for cogroup.
- groupWith(RDD<Tuple2<K, W1>>, RDD<Tuple2<K, W2>>, RDD<Tuple2<K, W3>>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Alias for cogroup.
- gt(double) - 类 中的静态方法org.apache.spark.ml.param.ParamValidators
-
Check if value is greater than lowerBound
- gt(Object) - 类 中的方法org.apache.spark.sql.Column
-
Greater than.
- gtEq(double) - 类 中的静态方法org.apache.spark.ml.param.ParamValidators
-
Check if value is greater than or equal to lowerBound
- guard(Function0<Parsers.Parser<T>>) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaParser
-
- L1Updater - org.apache.spark.mllib.optimization中的类
-
:: DeveloperApi ::
Updater for L1 regularized problems.
- L1Updater() - 类 的构造器org.apache.spark.mllib.optimization.L1Updater
-
- label() - 类 中的方法org.apache.spark.ml.feature.LabeledPoint
-
- label() - 类 中的方法org.apache.spark.mllib.regression.LabeledPoint
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- labelCol() - 接口 中的方法org.apache.spark.ml.classification.LogisticRegressionSummary
-
Field in "predictions" which gives the true label of each instance (if available).
- labelCol() - 类 中的方法org.apache.spark.ml.classification.LogisticRegressionSummaryImpl
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.classification.OneVsRest
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.classification.OneVsRestModel
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.evaluation.RegressionEvaluator
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelector
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.feature.RFormula
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.feature.RFormulaModel
-
- labelCol() - 接口 中的方法org.apache.spark.ml.param.shared.HasLabelCol
-
Param for label column name.
- labelCol() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegression
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegressionModel
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- labelCol() - 类 中的方法org.apache.spark.ml.regression.LinearRegressionSummary
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- labelCol() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- LabelConverter - org.apache.spark.ml.classification中的类
-
Label to vector converter.
- LabelConverter() - 类 的构造器org.apache.spark.ml.classification.LabelConverter
-
- LabeledPoint - org.apache.spark.ml.feature中的类
-
Class that represents the features and label of a data point.
- LabeledPoint(double, Vector) - 类 的构造器org.apache.spark.ml.feature.LabeledPoint
-
- LabeledPoint - org.apache.spark.mllib.regression中的类
-
Class that represents the features and labels of a data point.
- LabeledPoint(double, Vector) - 类 的构造器org.apache.spark.mllib.regression.LabeledPoint
-
- LabelPropagation - org.apache.spark.graphx.lib中的类
-
Label Propagation algorithm.
- LabelPropagation() - 类 的构造器org.apache.spark.graphx.lib.LabelPropagation
-
- labels() - 接口 中的方法org.apache.spark.ml.classification.LogisticRegressionSummary
-
Returns the sequence of labels in ascending order.
- labels() - 类 中的方法org.apache.spark.ml.feature.IndexToString
-
Optional param for array of labels specifying index-string mapping.
- labels() - 类 中的方法org.apache.spark.ml.feature.StringIndexerModel
-
- labels() - 类 中的方法org.apache.spark.mllib.classification.NaiveBayesModel
-
- labels() - 类 中的方法org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV1_0$.Data
-
- labels() - 类 中的方法org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV2_0$.Data
-
- labels() - 类 中的方法org.apache.spark.mllib.evaluation.MulticlassMetrics
-
Returns the sequence of labels in ascending order
- labels() - 类 中的方法org.apache.spark.mllib.evaluation.MultilabelMetrics
-
Returns the sequence of labels in ascending order
- lag(Column, int) - 类 中的静态方法org.apache.spark.sql.functions
-
Window function: returns the value that is offset rows before the current row, and
null if there is less than offset rows before the current row.
- lag(String, int) - 类 中的静态方法org.apache.spark.sql.functions
-
Window function: returns the value that is offset rows before the current row, and
null if there is less than offset rows before the current row.
- lag(String, int, Object) - 类 中的静态方法org.apache.spark.sql.functions
-
Window function: returns the value that is offset rows before the current row, and
defaultValue if there is less than offset rows before the current row.
- lag(Column, int, Object) - 类 中的静态方法org.apache.spark.sql.functions
-
Window function: returns the value that is offset rows before the current row, and
defaultValue if there is less than offset rows before the current row.
- LassoModel - org.apache.spark.mllib.regression中的类
-
Regression model trained using Lasso.
- LassoModel(Vector, double) - 类 的构造器org.apache.spark.mllib.regression.LassoModel
-
- LassoWithSGD - org.apache.spark.mllib.regression中的类
-
Train a regression model with L1-regularization using Stochastic Gradient Descent.
- LassoWithSGD() - 类 的构造器org.apache.spark.mllib.regression.LassoWithSGD
-
- last(Column, boolean) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the last value in a group.
- last(String, boolean) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the last value of the column in a group.
- last(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the last value in a group.
- last(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the last value of the column in a group.
- last() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- last(Object) - 类 中的方法org.apache.spark.util.kvstore.KVStoreView
-
Stops iteration at the given value of the chosen index (inclusive).
- last_day(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Given a date column, returns the last day of the month which the given date belongs to.
- lastDir() - 类 中的方法org.apache.spark.mllib.optimization.NNLS.Workspace
-
- lastError() - 类 中的方法org.apache.spark.status.api.v1.streaming.ReceiverInfo
-
- lastError() - 类 中的方法org.apache.spark.streaming.scheduler.ReceiverInfo
-
- lastErrorMessage() - 类 中的方法org.apache.spark.status.api.v1.streaming.ReceiverInfo
-
- lastErrorMessage() - 类 中的方法org.apache.spark.streaming.scheduler.ReceiverInfo
-
- lastErrorTime() - 类 中的方法org.apache.spark.status.api.v1.streaming.ReceiverInfo
-
- lastErrorTime() - 类 中的方法org.apache.spark.streaming.scheduler.ReceiverInfo
-
- lastIndexOf(B) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- lastIndexOf(B, int) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- lastIndexOfSlice(GenSeq<B>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- lastIndexOfSlice(GenSeq<B>, int) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- lastIndexWhere(Function1<A, Object>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- lastIndexWhere(Function1<A, Object>, int) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- lastOption() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- lastProgress() - 接口 中的方法org.apache.spark.sql.streaming.StreamingQuery
-
- lastStageNameAndDescription(org.apache.spark.status.AppStatusStore, JobData) - 类 中的静态方法org.apache.spark.ui.jobs.ApiHelper
-
- lastUpdate() - 类 中的方法org.apache.spark.status.LiveRDDDistribution
-
- lastUpdated() - 类 中的方法org.apache.spark.status.api.v1.ApplicationAttemptInfo
-
- Latest() - 类 的构造器org.apache.spark.streaming.kinesis.KinesisInitialPositions.Latest
-
- latestModel() - 类 中的方法org.apache.spark.mllib.clustering.StreamingKMeans
-
Return the latest model.
- latestModel() - 类 中的方法org.apache.spark.mllib.regression.StreamingLinearAlgorithm
-
Return the latest model.
- launch() - 类 中的方法org.apache.spark.launcher.SparkLauncher
-
Launches a sub-process that will start the configured Spark application.
- LAUNCH_TIME() - 类 中的静态方法org.apache.spark.status.TaskIndexNames
-
- LAUNCHING() - 类 中的静态方法org.apache.spark.TaskState
-
- LaunchTask(org.apache.spark.util.SerializableBuffer) - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.LaunchTask
-
- LaunchTask$() - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.LaunchTask$
-
- launchTime() - 类 中的方法org.apache.spark.scheduler.TaskInfo
-
- launchTime() - 类 中的方法org.apache.spark.status.api.v1.TaskData
-
- layers() - 类 中的方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- layers() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- LBFGS - org.apache.spark.mllib.optimization中的类
-
:: DeveloperApi ::
Class used to solve an optimization problem using Limited-memory BFGS.
- LBFGS(Gradient, Updater) - 类 的构造器org.apache.spark.mllib.optimization.LBFGS
-
- LDA - org.apache.spark.ml.clustering中的类
-
Latent Dirichlet Allocation (LDA), a topic model designed for text documents.
- LDA(String) - 类 的构造器org.apache.spark.ml.clustering.LDA
-
- LDA() - 类 的构造器org.apache.spark.ml.clustering.LDA
-
- LDA - org.apache.spark.mllib.clustering中的类
-
Latent Dirichlet Allocation (LDA), a topic model designed for text documents.
- LDA() - 类 的构造器org.apache.spark.mllib.clustering.LDA
-
Constructs a LDA instance with default parameters.
- LDAModel - org.apache.spark.ml.clustering中的类
-
- LDAModel - org.apache.spark.mllib.clustering中的类
-
Latent Dirichlet Allocation (LDA) model.
- LDAOptimizer - org.apache.spark.mllib.clustering中的接口
-
:: DeveloperApi ::
An LDAOptimizer specifies which optimization/learning/inference algorithm to use, and it can
hold optimizer-specific parameters for users to set.
- LDAUtils - org.apache.spark.mllib.clustering中的类
-
Utility methods for LDA.
- LDAUtils() - 类 的构造器org.apache.spark.mllib.clustering.LDAUtils
-
- lead(String, int) - 类 中的静态方法org.apache.spark.sql.functions
-
Window function: returns the value that is offset rows after the current row, and
null if there is less than offset rows after the current row.
- lead(Column, int) - 类 中的静态方法org.apache.spark.sql.functions
-
Window function: returns the value that is offset rows after the current row, and
null if there is less than offset rows after the current row.
- lead(String, int, Object) - 类 中的静态方法org.apache.spark.sql.functions
-
Window function: returns the value that is offset rows after the current row, and
defaultValue if there is less than offset rows after the current row.
- lead(Column, int, Object) - 类 中的静态方法org.apache.spark.sql.functions
-
Window function: returns the value that is offset rows after the current row, and
defaultValue if there is less than offset rows after the current row.
- LeafNode - org.apache.spark.ml.tree中的类
-
Decision tree leaf node.
- learningDecay() - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- learningDecay() - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- learningDecay() - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- learningOffset() - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- learningOffset() - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- learningOffset() - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- learningRate() - 类 中的方法org.apache.spark.mllib.tree.configuration.BoostingStrategy
-
- least(Column...) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the least value of the list of values, skipping null values.
- least(String, String...) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the least value of the list of column names, skipping null values.
- least(Seq<Column>) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the least value of the list of values, skipping null values.
- least(String, Seq<String>) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the least value of the list of column names, skipping null values.
- LeastSquaresGradient - org.apache.spark.mllib.optimization中的类
-
:: DeveloperApi ::
Compute gradient and loss for a Least-squared loss function, as used in linear regression.
- LeastSquaresGradient() - 类 的构造器org.apache.spark.mllib.optimization.LeastSquaresGradient
-
- left() - 类 中的方法org.apache.spark.sql.sources.And
-
- left() - 类 中的方法org.apache.spark.sql.sources.Or
-
- leftCategories() - 类 中的方法org.apache.spark.ml.tree.CategoricalSplit
-
Get sorted categories which split to the left
- leftCategoriesOrThreshold() - 类 中的方法org.apache.spark.ml.tree.DecisionTreeModelReadWrite.SplitData
-
- leftChild() - 类 中的方法org.apache.spark.ml.tree.DecisionTreeModelReadWrite.NodeData
-
- leftChild() - 类 中的方法org.apache.spark.ml.tree.InternalNode
-
- leftChildIndex(int) - 类 中的静态方法org.apache.spark.mllib.tree.model.Node
-
Return the index of the left child of this node.
- leftImpurity() - 类 中的方法org.apache.spark.mllib.tree.model.InformationGainStats
-
- leftJoin(RDD<Tuple2<Object, VD2>>, Function3<Object, VD, Option<VD2>, VD3>, ClassTag<VD2>, ClassTag<VD3>) - 类 中的方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- leftJoin(RDD<Tuple2<Object, VD2>>, Function3<Object, VD, Option<VD2>, VD3>, ClassTag<VD2>, ClassTag<VD3>) - 类 中的方法org.apache.spark.graphx.VertexRDD
-
Left joins this VertexRDD with an RDD containing vertex attribute pairs.
- leftNode() - 类 中的方法org.apache.spark.mllib.tree.model.Node
-
- leftNodeId() - 类 中的方法org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.NodeData
-
- leftOuterJoin(JavaPairRDD<K, W>, Partitioner) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Perform a left outer join of this and other.
- leftOuterJoin(JavaPairRDD<K, W>) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Perform a left outer join of this and other.
- leftOuterJoin(JavaPairRDD<K, W>, int) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Perform a left outer join of this and other.
- leftOuterJoin(RDD<Tuple2<K, W>>, Partitioner) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Perform a left outer join of this and other.
- leftOuterJoin(RDD<Tuple2<K, W>>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Perform a left outer join of this and other.
- leftOuterJoin(RDD<Tuple2<K, W>>, int) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Perform a left outer join of this and other.
- leftOuterJoin(JavaPairDStream<K, W>) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying 'left outer join' between RDDs of this DStream and
other DStream.
- leftOuterJoin(JavaPairDStream<K, W>, int) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying 'left outer join' between RDDs of this DStream and
other DStream.
- leftOuterJoin(JavaPairDStream<K, W>, Partitioner) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying 'left outer join' between RDDs of this DStream and
other DStream.
- leftOuterJoin(JavaPairDStream<K, W>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- leftOuterJoin(JavaPairDStream<K, W>, int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- leftOuterJoin(JavaPairDStream<K, W>, Partitioner) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- leftOuterJoin(JavaPairDStream<K, W>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- leftOuterJoin(JavaPairDStream<K, W>, int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- leftOuterJoin(JavaPairDStream<K, W>, Partitioner) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- leftOuterJoin(DStream<Tuple2<K, W>>, ClassTag<W>) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying 'left outer join' between RDDs of this DStream and
other DStream.
- leftOuterJoin(DStream<Tuple2<K, W>>, int, ClassTag<W>) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying 'left outer join' between RDDs of this DStream and
other DStream.
- leftOuterJoin(DStream<Tuple2<K, W>>, Partitioner, ClassTag<W>) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying 'left outer join' between RDDs of this DStream and
other DStream.
- leftPredict() - 类 中的方法org.apache.spark.mllib.tree.model.InformationGainStats
-
- leftZipJoin(VertexRDD<VD2>, Function3<Object, VD, Option<VD2>, VD3>, ClassTag<VD2>, ClassTag<VD3>) - 类 中的方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- leftZipJoin(VertexRDD<VD2>, Function3<Object, VD, Option<VD2>, VD3>, ClassTag<VD2>, ClassTag<VD3>) - 类 中的方法org.apache.spark.graphx.VertexRDD
-
Left joins this RDD with another VertexRDD with the same index.
- LegacyAccumulatorWrapper<R,T> - org.apache.spark.util中的类
-
- LegacyAccumulatorWrapper(R, AccumulableParam<R, T>) - 类 的构造器org.apache.spark.util.LegacyAccumulatorWrapper
-
- length() - 类 中的方法org.apache.spark.scheduler.SplitInfo
-
- length(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Computes the character length of a given string or number of bytes of a binary string.
- length() - 接口 中的方法org.apache.spark.sql.Row
-
Number of elements in the Row.
- length() - 类 中的方法org.apache.spark.sql.types.CharType
-
- length() - 类 中的方法org.apache.spark.sql.types.HiveStringType
-
- length() - 类 中的方法org.apache.spark.sql.types.StructType
-
- length() - 类 中的方法org.apache.spark.sql.types.VarcharType
-
- length() - 类 中的方法org.apache.spark.status.RDDPartitionSeq
-
- lengthCompare(int) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- leq(Object) - 类 中的方法org.apache.spark.sql.Column
-
Less than or equal to.
- less(Duration) - 类 中的方法org.apache.spark.streaming.Duration
-
- less(Time) - 类 中的方法org.apache.spark.streaming.Time
-
- lessEq(Duration) - 类 中的方法org.apache.spark.streaming.Duration
-
- lessEq(Time) - 类 中的方法org.apache.spark.streaming.Time
-
- LessThan - org.apache.spark.sql.sources中的类
-
A filter that evaluates to true iff the attribute evaluates to a value
less than value.
- LessThan(String, Object) - 类 的构造器org.apache.spark.sql.sources.LessThan
-
- LessThanOrEqual - org.apache.spark.sql.sources中的类
-
A filter that evaluates to true iff the attribute evaluates to a value
less than or equal to value.
- LessThanOrEqual(String, Object) - 类 的构造器org.apache.spark.sql.sources.LessThanOrEqual
-
- LevelDB - org.apache.spark.util.kvstore中的类
-
Implementation of KVStore that uses LevelDB as the underlying data store.
- LevelDB(File) - 类 的构造器org.apache.spark.util.kvstore.LevelDB
-
- LevelDB(File, KVStoreSerializer) - 类 的构造器org.apache.spark.util.kvstore.LevelDB
-
- LevelDB.TypeAliases - org.apache.spark.util.kvstore中的类
-
Needs to be public for Jackson.
- levenshtein(Column, Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Computes the Levenshtein distance of the two given string columns.
- libraryPathEnvName() - 类 中的静态方法org.apache.spark.util.Utils
-
Return the current system LD_LIBRARY_PATH name
- libraryPathEnvPrefix(Seq<String>) - 类 中的静态方法org.apache.spark.util.Utils
-
Return the prefix of a command that appends the given library paths to the
system-specific library path environment variable.
- LibSVMDataSource - org.apache.spark.ml.source.libsvm中的类
-
libsvm package implements Spark SQL data source API for loading LIBSVM data as DataFrame.
- lift() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- like(String) - 类 中的方法org.apache.spark.sql.Column
-
SQL like expression.
- limit(int) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new Dataset by taking the first n rows.
- line() - 异常错误 中的方法org.apache.spark.sql.AnalysisException
-
- LinearDataGenerator - org.apache.spark.mllib.util中的类
-
:: DeveloperApi ::
Generate sample data used for Linear Data.
- LinearDataGenerator() - 类 的构造器org.apache.spark.mllib.util.LinearDataGenerator
-
- LinearRegression - org.apache.spark.ml.regression中的类
-
Linear regression.
- LinearRegression(String) - 类 的构造器org.apache.spark.ml.regression.LinearRegression
-
- LinearRegression() - 类 的构造器org.apache.spark.ml.regression.LinearRegression
-
- LinearRegressionModel - org.apache.spark.ml.regression中的类
-
- LinearRegressionModel - org.apache.spark.mllib.regression中的类
-
Regression model trained using LinearRegression.
- LinearRegressionModel(Vector, double) - 类 的构造器org.apache.spark.mllib.regression.LinearRegressionModel
-
- LinearRegressionSummary - org.apache.spark.ml.regression中的类
-
:: Experimental ::
Linear regression results evaluated on a dataset.
- LinearRegressionTrainingSummary - org.apache.spark.ml.regression中的类
-
:: Experimental ::
Linear regression training results.
- LinearRegressionWithSGD - org.apache.spark.mllib.regression中的类
-
Train a linear regression model with no regularization using Stochastic Gradient Descent.
- LinearRegressionWithSGD() - 类 的构造器org.apache.spark.mllib.regression.LinearRegressionWithSGD
-
- LinearSVC - org.apache.spark.ml.classification中的类
-
:: Experimental ::
Linear SVM Classifier
This binary classifier optimizes the Hinge Loss using the OWLQN optimizer.
- LinearSVC(String) - 类 的构造器org.apache.spark.ml.classification.LinearSVC
-
- LinearSVC() - 类 的构造器org.apache.spark.ml.classification.LinearSVC
-
- LinearSVCModel - org.apache.spark.ml.classification中的类
-
:: Experimental ::
Linear SVM Model trained by
LinearSVC
- link(double) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.CLogLog$
-
- link(double) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.Identity$
-
- link(double) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.Inverse$
-
- link() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- link(double) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.Log$
-
- link(double) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.Logit$
-
- link(double) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.Probit$
-
- link(double) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.Sqrt$
-
- link() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- Link$() - 类 的构造器org.apache.spark.ml.regression.GeneralizedLinearRegression.Link$
-
- linkPower() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- linkPower() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- linkPredictionCol() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- linkPredictionCol() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- listColumns(String) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
Returns a list of columns for the given table/view or temporary view.
- listColumns(String, String) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
Returns a list of columns for the given table/view in the specified database.
- listDatabases() - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
Returns a list of databases available across all sessions.
- listenerManager() - 类 中的方法org.apache.spark.sql.SparkSession
-
- listenerManager() - 类 中的方法org.apache.spark.sql.SQLContext
-
- listFiles() - 类 中的方法org.apache.spark.SparkContext
-
Returns a list of file paths that are added to resources.
- listFunctions() - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
Returns a list of functions registered in the current database.
- listFunctions(String) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
Returns a list of functions registered in the specified database.
- listingTable(Seq<String>, Function1<T, Seq<Node>>, Iterable<T>, boolean, Option<String>, Seq<String>, boolean, boolean) - 类 中的静态方法org.apache.spark.ui.UIUtils
-
Returns an HTML table constructed by generating a row for each object in a sequence.
- listJars() - 类 中的方法org.apache.spark.SparkContext
-
Returns a list of jar files that are added to resources.
- listOrcFiles(String, Configuration) - 类 中的静态方法org.apache.spark.sql.hive.orc.OrcFileOperator
-
- listTables() - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
Returns a list of tables/views in the current database.
- listTables(String) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
Returns a list of tables/views in the specified database.
- lit(Object) - 类 中的静态方法org.apache.spark.sql.functions
-
Creates a
Column of literal value.
- literal(String) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaParser
-
- LIVE_ENTITY_UPDATE_PERIOD() - 类 中的静态方法org.apache.spark.status.config
-
- LiveEntityHelpers - org.apache.spark.status中的类
-
- LiveEntityHelpers() - 类 的构造器org.apache.spark.status.LiveEntityHelpers
-
- LiveExecutor - org.apache.spark.status中的类
-
- LiveExecutor(String, long) - 类 的构造器org.apache.spark.status.LiveExecutor
-
- LiveExecutorStageSummary - org.apache.spark.status中的类
-
- LiveExecutorStageSummary(int, int, String) - 类 的构造器org.apache.spark.status.LiveExecutorStageSummary
-
- LiveJob - org.apache.spark.status中的类
-
- LiveJob(int, String, Option<Date>, Seq<Object>, Option<String>, int) - 类 的构造器org.apache.spark.status.LiveJob
-
- LiveRDD - org.apache.spark.status中的类
-
- LiveRDD(RDDInfo) - 类 的构造器org.apache.spark.status.LiveRDD
-
- LiveRDDDistribution - org.apache.spark.status中的类
-
- LiveRDDDistribution(LiveExecutor) - 类 的构造器org.apache.spark.status.LiveRDDDistribution
-
- LiveRDDPartition - org.apache.spark.status中的类
-
- LiveRDDPartition(String) - 类 的构造器org.apache.spark.status.LiveRDDPartition
-
- LiveStage - org.apache.spark.status中的类
-
- LiveStage() - 类 的构造器org.apache.spark.status.LiveStage
-
- LiveTask - org.apache.spark.status中的类
-
- LiveTask(TaskInfo, int, int, Option<Object>) - 类 的构造器org.apache.spark.status.LiveTask
-
- load(String) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- load(String) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- load(String) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- load(String) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- load(String) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- load(String) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- load(String) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.classification.OneVsRest
-
- load(String) - 类 中的静态方法org.apache.spark.ml.classification.OneVsRestModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- load(String) - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeans
-
- load(String) - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixture
-
- load(String) - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.clustering.KMeans
-
- load(String) - 类 中的静态方法org.apache.spark.ml.clustering.KMeansModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- load(String) - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
-
- load(String) - 类 中的静态方法org.apache.spark.ml.evaluation.ClusteringEvaluator
-
- load(String) - 类 中的静态方法org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- load(String) - 类 中的静态方法org.apache.spark.ml.evaluation.RegressionEvaluator
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.Binarizer
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.Bucketizer
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelector
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.ColumnPruner
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizer
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizerModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.DCT
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.ElementwiseProduct
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.FeatureHasher
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.HashingTF
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.IDF
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.IDFModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.Imputer
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.ImputerModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.IndexToString
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.Interaction
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.MaxAbsScaler
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.MaxAbsScalerModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSH
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSHModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScaler
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScalerModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.NGram
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.Normalizer
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoder
-
已过时。
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderEstimator
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.PCA
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.PCAModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.PolynomialExpansion
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.QuantileDiscretizer
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.RegexTokenizer
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.RFormula
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.SQLTransformer
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.StandardScaler
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.StandardScalerModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.StopWordsRemover
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.StringIndexer
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.StringIndexerModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.Tokenizer
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.VectorAssembler
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.VectorAttributeRewriter
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexer
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexerModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.VectorSizeHint
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.VectorSlicer
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.Word2Vec
-
- load(String) - 类 中的静态方法org.apache.spark.ml.feature.Word2VecModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowth
-
- load(String) - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowthModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.Pipeline
-
- load(String, SparkContext, String) - 类 中的方法org.apache.spark.ml.Pipeline.SharedReadWrite$
-
- load(String) - 类 中的静态方法org.apache.spark.ml.PipelineModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.r.RWrappers
-
- load(String) - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- load(String) - 类 中的静态方法org.apache.spark.ml.recommendation.ALSModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- load(String) - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- load(String) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- load(String) - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- load(String) - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegression
-
- load(String) - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegressionModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- load(String) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- load(String) - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidator
-
- load(String) - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidatorModel
-
- load(String) - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplit
-
- load(String) - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplitModel
-
- load(String) - 接口 中的方法org.apache.spark.ml.util.MLReadable
-
Reads an ML instance from the input path, a shortcut of read.load(path).
- load(String) - 类 中的方法org.apache.spark.ml.util.MLReader
-
Loads the ML component from the input path.
- load(SparkContext, String) - 类 中的静态方法org.apache.spark.mllib.classification.LogisticRegressionModel
-
- load(SparkContext, String) - 类 中的静态方法org.apache.spark.mllib.classification.NaiveBayesModel
-
- load(SparkContext, String) - 类 中的方法org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV1_0$
-
- load(SparkContext, String) - 类 中的方法org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV2_0$
-
- load(SparkContext, String) - 类 中的静态方法org.apache.spark.mllib.classification.SVMModel
-
- load(SparkContext, String) - 类 中的静态方法org.apache.spark.mllib.clustering.BisectingKMeansModel
-
- load(SparkContext, String, int) - 类 中的方法org.apache.spark.mllib.clustering.BisectingKMeansModel.SaveLoadV1_0$
-
- load(SparkContext, String) - 类 中的静态方法org.apache.spark.mllib.clustering.DistributedLDAModel
-
- load(SparkContext, String) - 类 中的静态方法org.apache.spark.mllib.clustering.GaussianMixtureModel
-
- load(SparkContext, String) - 类 中的静态方法org.apache.spark.mllib.clustering.KMeansModel
-
- load(SparkContext, String) - 类 中的方法org.apache.spark.mllib.clustering.KMeansModel.SaveLoadV1_0$
-
- load(SparkContext, String) - 类 中的静态方法org.apache.spark.mllib.clustering.LocalLDAModel
-
- load(SparkContext, String) - 类 中的静态方法org.apache.spark.mllib.clustering.PowerIterationClusteringModel
-
- load(SparkContext, String) - 类 中的方法org.apache.spark.mllib.clustering.PowerIterationClusteringModel.SaveLoadV1_0$
-
- load(SparkContext, String) - 类 中的静态方法org.apache.spark.mllib.feature.ChiSqSelectorModel
-
- load(SparkContext, String) - 类 中的方法org.apache.spark.mllib.feature.ChiSqSelectorModel.SaveLoadV1_0$
-
- load(SparkContext, String) - 类 中的静态方法org.apache.spark.mllib.feature.Word2VecModel
-
- load(SparkContext, String) - 类 中的静态方法org.apache.spark.mllib.fpm.FPGrowthModel
-
- load(SparkContext, String) - 类 中的方法org.apache.spark.mllib.fpm.FPGrowthModel.SaveLoadV1_0$
-
- load(SparkContext, String) - 类 中的静态方法org.apache.spark.mllib.fpm.PrefixSpanModel
-
- load(SparkContext, String) - 类 中的方法org.apache.spark.mllib.fpm.PrefixSpanModel.SaveLoadV1_0$
-
- load(SparkContext, String) - 类 中的静态方法org.apache.spark.mllib.recommendation.MatrixFactorizationModel
-
Load a model from the given path.
- load(SparkContext, String) - 类 中的方法org.apache.spark.mllib.recommendation.MatrixFactorizationModel.SaveLoadV1_0$
-
- load(SparkContext, String) - 类 中的静态方法org.apache.spark.mllib.regression.IsotonicRegressionModel
-
- load(SparkContext, String) - 类 中的静态方法org.apache.spark.mllib.regression.LassoModel
-
- load(SparkContext, String) - 类 中的静态方法org.apache.spark.mllib.regression.LinearRegressionModel
-
- load(SparkContext, String) - 类 中的静态方法org.apache.spark.mllib.regression.RidgeRegressionModel
-
- load(SparkContext, String) - 类 中的静态方法org.apache.spark.mllib.tree.model.DecisionTreeModel
-
- load(SparkContext, String, String, int) - 类 中的方法org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$
-
- load(SparkContext, String) - 类 中的静态方法org.apache.spark.mllib.tree.model.GradientBoostedTreesModel
-
- load(SparkContext, String) - 类 中的静态方法org.apache.spark.mllib.tree.model.RandomForestModel
-
- load(SparkContext, String) - 接口 中的方法org.apache.spark.mllib.util.Loader
-
Load a model from the given path.
- load(String...) - 类 中的方法org.apache.spark.sql.DataFrameReader
-
Loads input in as a DataFrame, for data sources that support multiple paths.
- load() - 类 中的方法org.apache.spark.sql.DataFrameReader
-
Loads input in as a DataFrame, for data sources that don't require a path (e.g. external
key-value stores).
- load(String) - 类 中的方法org.apache.spark.sql.DataFrameReader
-
Loads input in as a DataFrame, for data sources that require a path (e.g. data backed by
a local or distributed file system).
- load(Seq<String>) - 类 中的方法org.apache.spark.sql.DataFrameReader
-
Loads input in as a DataFrame, for data sources that support multiple paths.
- load(String) - 类 中的方法org.apache.spark.sql.SQLContext
-
- load(String, String) - 类 中的方法org.apache.spark.sql.SQLContext
-
- load(String, Map<String, String>) - 类 中的方法org.apache.spark.sql.SQLContext
-
- load(String, Map<String, String>) - 类 中的方法org.apache.spark.sql.SQLContext
-
- load(String, StructType, Map<String, String>) - 类 中的方法org.apache.spark.sql.SQLContext
-
- load(String, StructType, Map<String, String>) - 类 中的方法org.apache.spark.sql.SQLContext
-
- load() - 类 中的方法org.apache.spark.sql.streaming.DataStreamReader
-
Loads input data stream in as a DataFrame, for data streams that don't require a path
(e.g. external key-value stores).
- load(String) - 类 中的方法org.apache.spark.sql.streaming.DataStreamReader
-
Loads input in as a DataFrame, for data streams that read from some path.
- loadData(SparkContext, String, String) - 类 中的方法org.apache.spark.mllib.classification.impl.GLMClassificationModel.SaveLoadV1_0$
-
Helper method for loading GLM classification model data.
- loadData(SparkContext, String, String, int) - 类 中的方法org.apache.spark.mllib.regression.impl.GLMRegressionModel.SaveLoadV1_0$
-
Helper method for loading GLM regression model data.
- loadDefaultSparkProperties(SparkConf, String) - 类 中的静态方法org.apache.spark.util.Utils
-
Load default Spark properties from the given file.
- loadDefaultStopWords(String) - 类 中的静态方法org.apache.spark.ml.feature.StopWordsRemover
-
Loads the default stop words for the given language.
- Loader<M extends Saveable> - org.apache.spark.mllib.util中的接口
-
:: DeveloperApi ::
Trait for classes which can load models and transformers from files.
- loadExtensions(Class<T>, Seq<String>, SparkConf) - 类 中的静态方法org.apache.spark.util.Utils
-
Create instances of extension classes.
- loadImpl(String, SparkSession, String, String) - 类 中的静态方法org.apache.spark.ml.tree.EnsembleModelReadWrite
-
Helper method for loading a tree ensemble from disk.
- loadImpl(Dataset<Row>, Item, ClassTag<Item>) - 类 中的方法org.apache.spark.mllib.fpm.FPGrowthModel.SaveLoadV1_0$
-
- loadImpl(Dataset<Row>, Item, ClassTag<Item>) - 类 中的方法org.apache.spark.mllib.fpm.PrefixSpanModel.SaveLoadV1_0$
-
- loadLabeledPoints(SparkContext, String, int) - 类 中的静态方法org.apache.spark.mllib.util.MLUtils
-
Loads labeled points saved using RDD[LabeledPoint].saveAsTextFile.
- loadLabeledPoints(SparkContext, String) - 类 中的静态方法org.apache.spark.mllib.util.MLUtils
-
Loads labeled points saved using RDD[LabeledPoint].saveAsTextFile with the default number of
partitions.
- loadLibSVMFile(SparkContext, String, int, int) - 类 中的静态方法org.apache.spark.mllib.util.MLUtils
-
Loads labeled data in the LIBSVM format into an RDD[LabeledPoint].
- loadLibSVMFile(SparkContext, String, int) - 类 中的静态方法org.apache.spark.mllib.util.MLUtils
-
Loads labeled data in the LIBSVM format into an RDD[LabeledPoint], with the default number of
partitions.
- loadLibSVMFile(SparkContext, String) - 类 中的静态方法org.apache.spark.mllib.util.MLUtils
-
Loads binary labeled data in the LIBSVM format into an RDD[LabeledPoint], with number of
features determined automatically and the default number of partitions.
- loadTreeNodes(String, org.apache.spark.ml.util.DefaultParamsReader.Metadata, SparkSession) - 类 中的静态方法org.apache.spark.ml.tree.DecisionTreeModelReadWrite
-
Load a decision tree from a file.
- loadVectors(SparkContext, String, int) - 类 中的静态方法org.apache.spark.mllib.util.MLUtils
-
Loads vectors saved using RDD[Vector].saveAsTextFile.
- loadVectors(SparkContext, String) - 类 中的静态方法org.apache.spark.mllib.util.MLUtils
-
Loads vectors saved using RDD[Vector].saveAsTextFile with the default number of partitions.
- LOCAL_BLOCKS_FETCHED() - 类 中的方法org.apache.spark.InternalAccumulator.shuffleRead$
-
- LOCAL_BYTES_READ() - 类 中的方法org.apache.spark.InternalAccumulator.shuffleRead$
-
- LOCAL_CLUSTER_REGEX() - 类 中的静态方法org.apache.spark.SparkMasterRegex
-
- LOCAL_N_FAILURES_REGEX() - 类 中的静态方法org.apache.spark.SparkMasterRegex
-
- LOCAL_N_REGEX() - 类 中的静态方法org.apache.spark.SparkMasterRegex
-
- localBlocksFetched() - 类 中的方法org.apache.spark.status.api.v1.ShuffleReadMetricDistributions
-
- localBlocksFetched() - 类 中的方法org.apache.spark.status.api.v1.ShuffleReadMetrics
-
- localBytesRead() - 类 中的方法org.apache.spark.status.api.v1.ShuffleReadMetrics
-
- localCanonicalHostName() - 类 中的静态方法org.apache.spark.util.Utils
-
Get the local machine's FQDN.
- localCheckpoint() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- localCheckpoint() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- localCheckpoint() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- localCheckpoint() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- localCheckpoint() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- localCheckpoint() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- localCheckpoint() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- localCheckpoint() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- localCheckpoint() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- localCheckpoint() - 类 中的方法org.apache.spark.rdd.RDD
-
Mark this RDD for local checkpointing using Spark's existing caching layer.
- localCheckpoint() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- localCheckpoint() - 类 中的方法org.apache.spark.sql.Dataset
-
Eagerly locally checkpoints a Dataset and return the new Dataset.
- localCheckpoint(boolean) - 类 中的方法org.apache.spark.sql.Dataset
-
Locally checkpoints a Dataset and return the new Dataset.
- localHostName() - 类 中的静态方法org.apache.spark.util.Utils
-
Get the local machine's hostname.
- localHostNameForURI() - 类 中的静态方法org.apache.spark.util.Utils
-
Get the local machine's URI.
- LOCALITY() - 类 中的静态方法org.apache.spark.status.TaskIndexNames
-
- localityAwareTasks() - 类 中的方法org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RequestExecutors
-
- localitySummary() - 类 中的方法org.apache.spark.status.LiveStage
-
- LocalKMeans - org.apache.spark.mllib.clustering中的类
-
An utility object to run K-means locally.
- LocalKMeans() - 类 的构造器org.apache.spark.mllib.clustering.LocalKMeans
-
- LocalLDAModel - org.apache.spark.ml.clustering中的类
-
Local (non-distributed) model fitted by
LDA.
- LocalLDAModel - org.apache.spark.mllib.clustering中的类
-
Local LDA model.
- localSeqToDatasetHolder(Seq<T>, Encoder<T>) - 类 中的方法org.apache.spark.sql.SQLImplicits
-
Creates a
Dataset from a local Seq.
- localSparkRPackagePath() - 类 中的静态方法org.apache.spark.api.r.RUtils
-
Get the SparkR package path in the local spark distribution.
- localValue() - 类 中的方法org.apache.spark.Accumulable
-
已过时。
Get the current value of this accumulator from within a task.
- localValue() - 类 中的静态方法org.apache.spark.Accumulator
-
已过时。
- locate(String, Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Locate the position of the first occurrence of substr.
- locate(String, Column, int) - 类 中的静态方法org.apache.spark.sql.functions
-
Locate the position of the first occurrence of substr in a string column, after position pos.
- location() - 类 中的方法org.apache.spark.streaming.scheduler.ReceiverInfo
-
- location() - 类 中的方法org.apache.spark.ui.storage.ExecutorStreamSummary
-
- locations() - 类 中的方法org.apache.spark.storage.BlockManagerMessages.BlockLocationsAndStatus
-
- locationUri() - 类 中的方法org.apache.spark.sql.catalog.Database
-
- log() - 接口 中的方法org.apache.spark.internal.Logging
-
- log(Function0<Parsers.Parser<T>>, String) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaParser
-
- log(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Computes the natural logarithm of the given value.
- log(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Computes the natural logarithm of the given column.
- log(double, Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the first argument-base logarithm of the second argument.
- log(double, String) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the first argument-base logarithm of the second argument.
- Log$() - 类 的构造器org.apache.spark.ml.regression.GeneralizedLinearRegression.Log$
-
- log10(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Computes the logarithm of the given value in base 10.
- log10(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Computes the logarithm of the given value in base 10.
- log1p(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Computes the natural logarithm of the given value plus one.
- log1p(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Computes the natural logarithm of the given column plus one.
- log2(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Computes the logarithm of the given column in base 2.
- log2(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Computes the logarithm of the given value in base 2.
- log_() - 接口 中的方法org.apache.spark.internal.Logging
-
- logDebug(Function0<String>) - 接口 中的方法org.apache.spark.internal.Logging
-
- logDebug(Function0<String>, Throwable) - 接口 中的方法org.apache.spark.internal.Logging
-
- logDeprecationWarning(String) - 类 中的静态方法org.apache.spark.SparkConf
-
Logs a warning message if the given config key is deprecated.
- logError(Function0<String>) - 接口 中的方法org.apache.spark.internal.Logging
-
- logError(Function0<String>, Throwable) - 接口 中的方法org.apache.spark.internal.Logging
-
- logEvent() - 接口 中的方法org.apache.spark.scheduler.SparkListenerEvent
-
- Logging - org.apache.spark.internal中的接口
-
Utility trait for classes that want to log data.
- logInfo(Function0<String>) - 接口 中的方法org.apache.spark.internal.Logging
-
- logInfo(Function0<String>, Throwable) - 接口 中的方法org.apache.spark.internal.Logging
-
- LogisticGradient - org.apache.spark.mllib.optimization中的类
-
:: DeveloperApi ::
Compute gradient and loss for a multinomial logistic loss function, as used
in multi-class classification (it is also used in binary logistic regression).
- LogisticGradient(int) - 类 的构造器org.apache.spark.mllib.optimization.LogisticGradient
-
- LogisticGradient() - 类 的构造器org.apache.spark.mllib.optimization.LogisticGradient
-
- LogisticRegression - org.apache.spark.ml.classification中的类
-
Logistic regression.
- LogisticRegression(String) - 类 的构造器org.apache.spark.ml.classification.LogisticRegression
-
- LogisticRegression() - 类 的构造器org.apache.spark.ml.classification.LogisticRegression
-
- LogisticRegressionDataGenerator - org.apache.spark.mllib.util中的类
-
:: DeveloperApi ::
Generate test data for LogisticRegression.
- LogisticRegressionDataGenerator() - 类 的构造器org.apache.spark.mllib.util.LogisticRegressionDataGenerator
-
- LogisticRegressionModel - org.apache.spark.ml.classification中的类
-
- LogisticRegressionModel - org.apache.spark.mllib.classification中的类
-
Classification model trained using Multinomial/Binary Logistic Regression.
- LogisticRegressionModel(Vector, double, int, int) - 类 的构造器org.apache.spark.mllib.classification.LogisticRegressionModel
-
- LogisticRegressionModel(Vector, double) - 类 的构造器org.apache.spark.mllib.classification.LogisticRegressionModel
-
- LogisticRegressionSummary - org.apache.spark.ml.classification中的接口
-
:: Experimental ::
Abstraction for logistic regression results for a given model.
- LogisticRegressionSummaryImpl - org.apache.spark.ml.classification中的类
-
Multiclass logistic regression results for a given model.
- LogisticRegressionSummaryImpl(Dataset<Row>, String, String, String, String) - 类 的构造器org.apache.spark.ml.classification.LogisticRegressionSummaryImpl
-
- LogisticRegressionTrainingSummary - org.apache.spark.ml.classification中的接口
-
:: Experimental ::
Abstraction for multiclass logistic regression training results.
- LogisticRegressionTrainingSummaryImpl - org.apache.spark.ml.classification中的类
-
Multiclass logistic regression training results.
- LogisticRegressionTrainingSummaryImpl(Dataset<Row>, String, String, String, String, double[]) - 类 的构造器org.apache.spark.ml.classification.LogisticRegressionTrainingSummaryImpl
-
- LogisticRegressionWithLBFGS - org.apache.spark.mllib.classification中的类
-
Train a classification model for Multinomial/Binary Logistic Regression using
Limited-memory BFGS.
- LogisticRegressionWithLBFGS() - 类 的构造器org.apache.spark.mllib.classification.LogisticRegressionWithLBFGS
-
- LogisticRegressionWithSGD - org.apache.spark.mllib.classification中的类
-
Train a classification model for Binary Logistic Regression
using Stochastic Gradient Descent.
- LogisticRegressionWithSGD() - 类 的构造器org.apache.spark.mllib.classification.LogisticRegressionWithSGD
-
- Logit$() - 类 的构造器org.apache.spark.ml.regression.GeneralizedLinearRegression.Logit$
-
- logLikelihood(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- logLikelihood() - 类 中的方法org.apache.spark.ml.clustering.ExpectationAggregator
-
- logLikelihood() - 类 中的方法org.apache.spark.ml.clustering.GaussianMixtureSummary
-
- logLikelihood(Dataset<?>) - 类 中的方法org.apache.spark.ml.clustering.LDAModel
-
Calculates a lower bound on the log likelihood of the entire corpus.
- logLikelihood(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- logLikelihood() - 类 中的方法org.apache.spark.mllib.clustering.DistributedLDAModel
-
Log likelihood of the observed tokens in the training set,
given the current parameter estimates:
log P(docs | topics, topic distributions for docs, alpha, eta)
Note:
- This excludes the prior; for that, use logPrior
- logLikelihood() - 类 中的方法org.apache.spark.mllib.clustering.ExpectationSum
-
- logLikelihood(RDD<Tuple2<Object, Vector>>) - 类 中的方法org.apache.spark.mllib.clustering.LocalLDAModel
-
Calculates a lower bound on the log likelihood of the entire corpus.
- logLikelihood(JavaPairRDD<Long, Vector>) - 类 中的方法org.apache.spark.mllib.clustering.LocalLDAModel
-
Java-friendly version of logLikelihood
- LogLoss - org.apache.spark.mllib.tree.loss中的类
-
:: DeveloperApi ::
Class for log loss calculation (for classification).
- LogLoss() - 类 的构造器org.apache.spark.mllib.tree.loss.LogLoss
-
- logName() - 接口 中的方法org.apache.spark.internal.Logging
-
- LogNormalGenerator - org.apache.spark.mllib.random中的类
-
:: DeveloperApi ::
Generates i.i.d. samples from the log normal distribution with the
given mean and standard deviation.
- LogNormalGenerator(double, double) - 类 的构造器org.apache.spark.mllib.random.LogNormalGenerator
-
- logNormalGraph(SparkContext, int, int, double, double, long) - 类 中的静态方法org.apache.spark.graphx.util.GraphGenerators
-
Generate a graph whose vertex out degree distribution is log normal.
- logNormalJavaRDD(JavaSparkContext, double, double, long, int, long) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
Java-friendly version of RandomRDDs.logNormalRDD.
- logNormalJavaRDD(JavaSparkContext, double, double, long, int) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
RandomRDDs.logNormalJavaRDD with the default seed.
- logNormalJavaRDD(JavaSparkContext, double, double, long) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
RandomRDDs.logNormalJavaRDD with the default number of partitions and the default seed.
- logNormalJavaVectorRDD(JavaSparkContext, double, double, long, int, int, long) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
Java-friendly version of RandomRDDs.logNormalVectorRDD.
- logNormalJavaVectorRDD(JavaSparkContext, double, double, long, int, int) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
RandomRDDs.logNormalJavaVectorRDD with the default seed.
- logNormalJavaVectorRDD(JavaSparkContext, double, double, long, int) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
RandomRDDs.logNormalJavaVectorRDD with the default number of partitions and
the default seed.
- logNormalRDD(SparkContext, double, double, long, int, long) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
Generates an RDD comprised of i.i.d.
- logNormalVectorRDD(SparkContext, double, double, long, int, int, long) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
Generates an RDD[Vector] with vectors containing i.i.d.
- logpdf(Vector) - 类 中的方法org.apache.spark.ml.stat.distribution.MultivariateGaussian
-
Returns the log-density of this multivariate Gaussian at given point, x
- logpdf(Vector) - 类 中的方法org.apache.spark.mllib.stat.distribution.MultivariateGaussian
-
Returns the log-density of this multivariate Gaussian at given point, x
- logPerplexity(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- logPerplexity(Dataset<?>) - 类 中的方法org.apache.spark.ml.clustering.LDAModel
-
Calculate an upper bound on perplexity.
- logPerplexity(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- logPerplexity(RDD<Tuple2<Object, Vector>>) - 类 中的方法org.apache.spark.mllib.clustering.LocalLDAModel
-
Calculate an upper bound on perplexity.
- logPerplexity(JavaPairRDD<Long, Vector>) - 类 中的方法org.apache.spark.mllib.clustering.LocalLDAModel
-
Java-friendly version of logPerplexity
- logPrior() - 类 中的方法org.apache.spark.ml.clustering.DistributedLDAModel
-
Log probability of the current parameter estimate:
log P(topics, topic distributions for docs | Dirichlet hyperparameters)
- logPrior() - 类 中的方法org.apache.spark.mllib.clustering.DistributedLDAModel
-
Log probability of the current parameter estimate:
log P(topics, topic distributions for docs | alpha, eta)
- logStartFromJson(JsonAST.JValue) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- logStartToJson(SparkListenerLogStart) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- logTrace(Function0<String>) - 接口 中的方法org.apache.spark.internal.Logging
-
- logTrace(Function0<String>, Throwable) - 接口 中的方法org.apache.spark.internal.Logging
-
- logUncaughtExceptions(Function0<T>) - 类 中的静态方法org.apache.spark.util.Utils
-
Execute the given block, logging and re-throwing any uncaught exception.
- logUrlMap() - 类 中的方法org.apache.spark.scheduler.cluster.ExecutorInfo
-
- logUrls() - 类 中的方法org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RegisterExecutor
-
- logWarning(Function0<String>) - 接口 中的方法org.apache.spark.internal.Logging
-
- logWarning(Function0<String>, Throwable) - 接口 中的方法org.apache.spark.internal.Logging
-
- LONG() - 类 中的静态方法org.apache.spark.sql.Encoders
-
An encoder for nullable long type.
- longAccumulator() - 类 中的方法org.apache.spark.SparkContext
-
Create and register a long accumulator, which starts with 0 and accumulates inputs by add.
- longAccumulator(String) - 类 中的方法org.apache.spark.SparkContext
-
Create and register a long accumulator, which starts with 0 and accumulates inputs by add.
- LongAccumulator - org.apache.spark.util中的类
-
An
accumulator for computing sum, count, and average of 64-bit integers.
- LongAccumulator() - 类 的构造器org.apache.spark.util.LongAccumulator
-
- LongAccumulatorParam$() - 类 的构造器org.apache.spark.AccumulatorParam.LongAccumulatorParam$
-
已过时。
- longMetric(String) - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- LongParam - org.apache.spark.ml.param中的类
-
:: DeveloperApi ::
Specialized version of Param[Long] for Java.
- LongParam(String, String, String, Function1<Object, Object>) - 类 的构造器org.apache.spark.ml.param.LongParam
-
- LongParam(String, String, String) - 类 的构造器org.apache.spark.ml.param.LongParam
-
- LongParam(Identifiable, String, String, Function1<Object, Object>) - 类 的构造器org.apache.spark.ml.param.LongParam
-
- LongParam(Identifiable, String, String) - 类 的构造器org.apache.spark.ml.param.LongParam
-
- LongType - 类 中的静态变量org.apache.spark.sql.types.DataTypes
-
Gets the LongType object.
- LongType - org.apache.spark.sql.types中的类
-
The data type representing Long values.
- lookup(K) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Return the list of values in the RDD for key key.
- lookup(K) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Return the list of values in the RDD for key key.
- lookupRpcTimeout(SparkConf) - 类 中的静态方法org.apache.spark.util.RpcUtils
-
Returns the default Spark timeout to use for RPC remote endpoint lookup.
- loss() - 接口 中的方法org.apache.spark.ml.param.shared.HasLoss
-
Param for the loss function to be optimized.
- loss() - 类 中的方法org.apache.spark.ml.regression.AFTAggregator
-
- loss() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- loss() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- loss() - 类 中的方法org.apache.spark.mllib.tree.configuration.BoostingStrategy
-
- Loss - org.apache.spark.mllib.tree.loss中的接口
-
:: DeveloperApi ::
Trait for adding "pluggable" loss functions for the gradient boosting algorithm.
- Losses - org.apache.spark.mllib.tree.loss中的类
-
- Losses() - 类 的构造器org.apache.spark.mllib.tree.loss.Losses
-
- LossReasonPending - org.apache.spark.scheduler中的类
-
A loss reason that means we don't yet know why the executor exited.
- LossReasonPending() - 类 的构造器org.apache.spark.scheduler.LossReasonPending
-
- lossType() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- lossType() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- lossType() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- lossType() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- LOST() - 类 中的静态方法org.apache.spark.TaskState
-
- low() - 类 中的方法org.apache.spark.partial.BoundedDouble
-
- lower(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Converts a string column to lower case.
- lowerBoundsOnCoefficients() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- lowerBoundsOnCoefficients() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- lowerBoundsOnIntercepts() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- lowerBoundsOnIntercepts() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- LowPrioritySQLImplicits - org.apache.spark.sql中的接口
-
Lower priority implicit methods for converting Scala objects into
Datasets.
- lpad(Column, int, String) - 类 中的静态方法org.apache.spark.sql.functions
-
Left-pad the string column with pad to a length of len.
- lt(double) - 类 中的静态方法org.apache.spark.ml.param.ParamValidators
-
Check if value is less than upperBound
- lt(Object) - 类 中的方法org.apache.spark.sql.Column
-
Less than.
- ltEq(double) - 类 中的静态方法org.apache.spark.ml.param.ParamValidators
-
Check if value is less than or equal to upperBound
- ltrim(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Trim the spaces from left end for the specified string value.
- ltrim(Column, String) - 类 中的静态方法org.apache.spark.sql.functions
-
Trim the specified character string from left end for the specified string column.
- LZ4CompressionCodec - org.apache.spark.io中的类
-
- LZ4CompressionCodec(SparkConf) - 类 的构造器org.apache.spark.io.LZ4CompressionCodec
-
- LZFCompressionCodec - org.apache.spark.io中的类
-
- LZFCompressionCodec(SparkConf) - 类 的构造器org.apache.spark.io.LZFCompressionCodec
-
- main(String[]) - 类 中的静态方法org.apache.spark.ml.param.shared.SharedParamsCodeGen
-
- main(String[]) - 类 中的静态方法org.apache.spark.mllib.util.KMeansDataGenerator
-
- main(String[]) - 类 中的静态方法org.apache.spark.mllib.util.LinearDataGenerator
-
- main(String[]) - 类 中的静态方法org.apache.spark.mllib.util.LogisticRegressionDataGenerator
-
- main(String[]) - 类 中的静态方法org.apache.spark.mllib.util.MFDataGenerator
-
- main(String[]) - 类 中的静态方法org.apache.spark.mllib.util.SVMDataGenerator
-
- main(String[]) - 类 中的静态方法org.apache.spark.streaming.util.RawTextSender
-
- main(String[]) - 类 中的静态方法org.apache.spark.ui.UIWorkloadGenerator
-
- majorMinorVersion(String) - 类 中的静态方法org.apache.spark.util.VersionUtils
-
Given a Spark version string, return the (major version number, minor version number).
- majorVersion(String) - 类 中的静态方法org.apache.spark.util.VersionUtils
-
Given a Spark version string, return the major version number.
- makeBinarySearch(Ordering<K>, ClassTag<K>) - 类 中的静态方法org.apache.spark.util.CollectionsUtils
-
- makeCopy(Object[]) - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- makeCopy(Object[]) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- makeCopy(Object[]) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- makeCopy(Object[]) - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- makeDescription(String, String, boolean) - 类 中的静态方法org.apache.spark.ui.UIUtils
-
Returns HTML rendering of a job or stage description.
- makeDriverRef(String, SparkConf, org.apache.spark.rpc.RpcEnv) - 类 中的静态方法org.apache.spark.util.RpcUtils
-
Retrieve a RpcEndpointRef which is located in the driver via its name.
- makeHref(boolean, String, String) - 类 中的静态方法org.apache.spark.ui.UIUtils
-
Return the correct Href after checking if master is running in the
reverse proxy mode or not.
- makeProgressBar(int, int, int, int, Map<String, Object>, int) - 类 中的静态方法org.apache.spark.ui.UIUtils
-
- makeRDD(Seq<T>, int, ClassTag<T>) - 类 中的方法org.apache.spark.SparkContext
-
Distribute a local Scala collection to form an RDD.
- makeRDD(Seq<Tuple2<T, Seq<String>>>, ClassTag<T>) - 类 中的方法org.apache.spark.SparkContext
-
Distribute a local Scala collection to form an RDD, with one or more
location preferences (hostnames of Spark nodes) for each object.
- map(Function<T, R>) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- map(Function<T, R>) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- map(Function<T, R>) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- map(Function<T, R>) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Return a new RDD by applying a function to all elements of this RDD.
- map(Function1<T, U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- map(Function1<T, U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- map(Function1<T, U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- map(Function1<T, U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- map(Function1<T, U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- map(Function1<Object, Object>) - 接口 中的方法org.apache.spark.ml.linalg.Matrix
-
Map the values of this matrix using a function.
- map(Function1<Object, Object>) - 接口 中的方法org.apache.spark.mllib.linalg.Matrix
-
Map the values of this matrix using a function.
- map(Function1<R, T>) - 类 中的方法org.apache.spark.partial.PartialResult
-
Transform this PartialResult into a PartialResult of type T.
- map(Function1<T, U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- map(Function1<T, U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- map(Function1<T, U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- map(Function1<T, U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- map(Function1<T, U>, ClassTag<U>) - 类 中的方法org.apache.spark.rdd.RDD
-
Return a new RDD by applying a function to all elements of this RDD.
- map(Function1<T, U>, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- map(DataType, DataType) - 类 中的方法org.apache.spark.sql.ColumnName
-
Creates a new StructField of type map.
- map(MapType) - 类 中的方法org.apache.spark.sql.ColumnName
-
- map(Function1<T, U>, Encoder<U>) - 类 中的方法org.apache.spark.sql.Dataset
-
:: Experimental ::
(Scala-specific)
Returns a new Dataset that contains the result of applying func to each element.
- map(MapFunction<T, U>, Encoder<U>) - 类 中的方法org.apache.spark.sql.Dataset
-
:: Experimental ::
(Java-specific)
Returns a new Dataset that contains the result of applying func to each element.
- map(Column...) - 类 中的静态方法org.apache.spark.sql.functions
-
Creates a new map column.
- map(Seq<Column>) - 类 中的静态方法org.apache.spark.sql.functions
-
Creates a new map column.
- map(Function1<BaseType, A>) - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- map(Function1<BaseType, A>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- map(Function1<BaseType, A>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- map(Function1<BaseType, A>) - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- map(Function1<A, B>, CanBuildFrom<Repr, B, That>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- map(Function<T, U>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- map(Function<T, U>) - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return a new DStream by applying a function to all elements of this DStream.
- map(Function<T, U>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- map(Function<T, U>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- map(Function<T, U>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- map(Function<T, U>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- map(Function<T, U>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- map(Function1<T, U>, ClassTag<U>) - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Return a new DStream by applying a function to all elements of this DStream.
- map_keys(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns an unordered array containing the keys of the map.
- map_values(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns an unordered array containing the values of the map.
- mapAsSerializableJavaMap(Map<A, B>) - 类 中的静态方法org.apache.spark.api.java.JavaUtils
-
- mapChildren(Function1<BaseType, BaseType>) - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- mapChildren(Function1<BaseType, BaseType>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- mapChildren(Function1<BaseType, BaseType>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- mapChildren(Function1<BaseType, BaseType>) - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- mapEdgePartitions(Function2<Object, EdgePartition<ED, VD>, EdgePartition<ED2, VD2>>, ClassTag<ED2>, ClassTag<VD2>) - 类 中的方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- mapEdges(Function1<Edge<ED>, ED2>, ClassTag<ED2>) - 类 中的方法org.apache.spark.graphx.Graph
-
Transforms each edge attribute in the graph using the map function.
- mapEdges(Function2<Object, Iterator<Edge<ED>>, Iterator<ED2>>, ClassTag<ED2>) - 类 中的方法org.apache.spark.graphx.Graph
-
Transforms each edge attribute using the map function, passing it a whole partition at a
time.
- mapEdges(Function2<Object, Iterator<Edge<ED>>, Iterator<ED2>>, ClassTag<ED2>) - 类 中的方法org.apache.spark.graphx.impl.GraphImpl
-
- mapExpressions(Function1<Expression, Expression>) - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- mapExpressions(Function1<Expression, Expression>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- mapExpressions(Function1<Expression, Expression>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- mapExpressions(Function1<Expression, Expression>) - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- mapFromJson(JsonAST.JValue) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
-------------------------------- *
Util JSON deserialization methods |
- MapFunction<T,U> - org.apache.spark.api.java.function中的接口
-
Base interface for a map function used in Dataset's map function.
- mapGroups(Function2<K, Iterator<V>, U>, Encoder<U>) - 类 中的方法org.apache.spark.sql.KeyValueGroupedDataset
-
(Scala-specific)
Applies the given function to each group of data.
- mapGroups(MapGroupsFunction<K, V, U>, Encoder<U>) - 类 中的方法org.apache.spark.sql.KeyValueGroupedDataset
-
(Java-specific)
Applies the given function to each group of data.
- MapGroupsFunction<K,V,R> - org.apache.spark.api.java.function中的接口
-
Base interface for a map function used in GroupedDataset's mapGroup function.
- mapGroupsWithState(Function3<K, Iterator<V>, GroupState<S>, U>, Encoder<S>, Encoder<U>) - 类 中的方法org.apache.spark.sql.KeyValueGroupedDataset
-
::Experimental::
(Scala-specific)
Applies the given function to each group of data, while maintaining a user-defined per-group
state.
- mapGroupsWithState(GroupStateTimeout, Function3<K, Iterator<V>, GroupState<S>, U>, Encoder<S>, Encoder<U>) - 类 中的方法org.apache.spark.sql.KeyValueGroupedDataset
-
::Experimental::
(Scala-specific)
Applies the given function to each group of data, while maintaining a user-defined per-group
state.
- mapGroupsWithState(MapGroupsWithStateFunction<K, V, S, U>, Encoder<S>, Encoder<U>) - 类 中的方法org.apache.spark.sql.KeyValueGroupedDataset
-
::Experimental::
(Java-specific)
Applies the given function to each group of data, while maintaining a user-defined per-group
state.
- mapGroupsWithState(MapGroupsWithStateFunction<K, V, S, U>, Encoder<S>, Encoder<U>, GroupStateTimeout) - 类 中的方法org.apache.spark.sql.KeyValueGroupedDataset
-
::Experimental::
(Java-specific)
Applies the given function to each group of data, while maintaining a user-defined per-group
state.
- MapGroupsWithStateFunction<K,V,S,R> - org.apache.spark.api.java.function中的接口
-
- mapId() - 类 中的方法org.apache.spark.FetchFailed
-
- mapId() - 类 中的方法org.apache.spark.storage.ShuffleBlockId
-
- mapId() - 类 中的方法org.apache.spark.storage.ShuffleDataBlockId
-
- mapId() - 类 中的方法org.apache.spark.storage.ShuffleIndexBlockId
-
- mapOutputTracker() - 类 中的方法org.apache.spark.SparkEnv
-
- mapPartitions(FlatMapFunction<Iterator<T>, U>) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- mapPartitions(FlatMapFunction<Iterator<T>, U>, boolean) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- mapPartitions(FlatMapFunction<Iterator<T>, U>) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- mapPartitions(FlatMapFunction<Iterator<T>, U>, boolean) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- mapPartitions(FlatMapFunction<Iterator<T>, U>) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- mapPartitions(FlatMapFunction<Iterator<T>, U>, boolean) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- mapPartitions(FlatMapFunction<Iterator<T>, U>) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Return a new RDD by applying a function to each partition of this RDD.
- mapPartitions(FlatMapFunction<Iterator<T>, U>, boolean) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Return a new RDD by applying a function to each partition of this RDD.
- mapPartitions(Function1<Iterator<T>, Iterator<U>>, boolean, ClassTag<U>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- mapPartitions(Function1<Iterator<T>, Iterator<U>>, boolean, ClassTag<U>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- mapPartitions(Function1<Iterator<T>, Iterator<U>>, boolean, ClassTag<U>) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- mapPartitions(Function1<Iterator<T>, Iterator<U>>, boolean, ClassTag<U>) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- mapPartitions(Function1<Iterator<T>, Iterator<U>>, boolean, ClassTag<U>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- mapPartitions(Function1<Iterator<T>, Iterator<U>>, boolean, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- mapPartitions(Function1<Iterator<T>, Iterator<U>>, boolean, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- mapPartitions(Function1<Iterator<T>, Iterator<U>>, boolean, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- mapPartitions(Function1<Iterator<T>, Iterator<U>>, boolean, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- mapPartitions(Function1<Iterator<T>, Iterator<U>>, boolean, ClassTag<U>) - 类 中的方法org.apache.spark.rdd.RDD
-
Return a new RDD by applying a function to each partition of this RDD.
- mapPartitions(Function1<Iterator<T>, Iterator<U>>, boolean, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- mapPartitions(Function1<Iterator<T>, Iterator<U>>, Encoder<U>) - 类 中的方法org.apache.spark.sql.Dataset
-
:: Experimental ::
(Scala-specific)
Returns a new Dataset that contains the result of applying func to each partition.
- mapPartitions(MapPartitionsFunction<T, U>, Encoder<U>) - 类 中的方法org.apache.spark.sql.Dataset
-
:: Experimental ::
(Java-specific)
Returns a new Dataset that contains the result of applying f to each partition.
- mapPartitions(FlatMapFunction<Iterator<T>, U>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- mapPartitions(FlatMapFunction<Iterator<T>, U>) - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return a new DStream in which each RDD is generated by applying mapPartitions() to each RDDs
of this DStream.
- mapPartitions(FlatMapFunction<Iterator<T>, U>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- mapPartitions(FlatMapFunction<Iterator<T>, U>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- mapPartitions(FlatMapFunction<Iterator<T>, U>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- mapPartitions(FlatMapFunction<Iterator<T>, U>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- mapPartitions(FlatMapFunction<Iterator<T>, U>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- mapPartitions(Function1<Iterator<T>, Iterator<U>>, boolean, ClassTag<U>) - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Return a new DStream in which each RDD is generated by applying mapPartitions() to each RDDs
of this DStream.
- mapPartitions$default$2() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- mapPartitions$default$2() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- mapPartitions$default$2() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- mapPartitions$default$2() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- mapPartitions$default$2() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- mapPartitions$default$2() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- mapPartitions$default$2() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- mapPartitions$default$2() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- mapPartitions$default$2() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- mapPartitions$default$2() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- MapPartitionsFunction<T,U> - org.apache.spark.api.java.function中的接口
-
Base interface for function used in Dataset's mapPartitions.
- mapPartitionsInternal$default$2() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- mapPartitionsInternal$default$2() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- mapPartitionsInternal$default$2() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- mapPartitionsInternal$default$2() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- mapPartitionsInternal$default$2() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- mapPartitionsInternal$default$2() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- mapPartitionsInternal$default$2() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- mapPartitionsInternal$default$2() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- mapPartitionsInternal$default$2() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- mapPartitionsInternal$default$2() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- mapPartitionsToDouble(DoubleFlatMapFunction<Iterator<T>>) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- mapPartitionsToDouble(DoubleFlatMapFunction<Iterator<T>>, boolean) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- mapPartitionsToDouble(DoubleFlatMapFunction<Iterator<T>>) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- mapPartitionsToDouble(DoubleFlatMapFunction<Iterator<T>>, boolean) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- mapPartitionsToDouble(DoubleFlatMapFunction<Iterator<T>>) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- mapPartitionsToDouble(DoubleFlatMapFunction<Iterator<T>>, boolean) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- mapPartitionsToDouble(DoubleFlatMapFunction<Iterator<T>>) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Return a new RDD by applying a function to each partition of this RDD.
- mapPartitionsToDouble(DoubleFlatMapFunction<Iterator<T>>, boolean) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Return a new RDD by applying a function to each partition of this RDD.
- mapPartitionsToPair(PairFlatMapFunction<Iterator<T>, K2, V2>) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- mapPartitionsToPair(PairFlatMapFunction<Iterator<T>, K2, V2>, boolean) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- mapPartitionsToPair(PairFlatMapFunction<Iterator<T>, K2, V2>) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- mapPartitionsToPair(PairFlatMapFunction<Iterator<T>, K2, V2>, boolean) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- mapPartitionsToPair(PairFlatMapFunction<Iterator<T>, K2, V2>) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- mapPartitionsToPair(PairFlatMapFunction<Iterator<T>, K2, V2>, boolean) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- mapPartitionsToPair(PairFlatMapFunction<Iterator<T>, K2, V2>) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Return a new RDD by applying a function to each partition of this RDD.
- mapPartitionsToPair(PairFlatMapFunction<Iterator<T>, K2, V2>, boolean) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Return a new RDD by applying a function to each partition of this RDD.
- mapPartitionsToPair(PairFlatMapFunction<Iterator<T>, K2, V2>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- mapPartitionsToPair(PairFlatMapFunction<Iterator<T>, K2, V2>) - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return a new DStream in which each RDD is generated by applying mapPartitions() to each RDDs
of this DStream.
- mapPartitionsToPair(PairFlatMapFunction<Iterator<T>, K2, V2>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- mapPartitionsToPair(PairFlatMapFunction<Iterator<T>, K2, V2>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- mapPartitionsToPair(PairFlatMapFunction<Iterator<T>, K2, V2>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- mapPartitionsToPair(PairFlatMapFunction<Iterator<T>, K2, V2>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- mapPartitionsToPair(PairFlatMapFunction<Iterator<T>, K2, V2>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- mapPartitionsWithIndex(Function2<Integer, Iterator<T>, Iterator<R>>, boolean) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- mapPartitionsWithIndex(Function2<Integer, Iterator<T>, Iterator<R>>, boolean) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- mapPartitionsWithIndex(Function2<Integer, Iterator<T>, Iterator<R>>, boolean) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- mapPartitionsWithIndex(Function2<Integer, Iterator<T>, Iterator<R>>, boolean) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Return a new RDD by applying a function to each partition of this RDD, while tracking the index
of the original partition.
- mapPartitionsWithIndex(Function2<Object, Iterator<T>, Iterator<U>>, boolean, ClassTag<U>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- mapPartitionsWithIndex(Function2<Object, Iterator<T>, Iterator<U>>, boolean, ClassTag<U>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- mapPartitionsWithIndex(Function2<Object, Iterator<T>, Iterator<U>>, boolean, ClassTag<U>) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- mapPartitionsWithIndex(Function2<Object, Iterator<T>, Iterator<U>>, boolean, ClassTag<U>) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- mapPartitionsWithIndex(Function2<Object, Iterator<T>, Iterator<U>>, boolean, ClassTag<U>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- mapPartitionsWithIndex(Function2<Object, Iterator<T>, Iterator<U>>, boolean, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- mapPartitionsWithIndex(Function2<Object, Iterator<T>, Iterator<U>>, boolean, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- mapPartitionsWithIndex(Function2<Object, Iterator<T>, Iterator<U>>, boolean, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- mapPartitionsWithIndex(Function2<Object, Iterator<T>, Iterator<U>>, boolean, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- mapPartitionsWithIndex(Function2<Object, Iterator<T>, Iterator<U>>, boolean, ClassTag<U>) - 类 中的方法org.apache.spark.rdd.RDD
-
Return a new RDD by applying a function to each partition of this RDD, while tracking the index
of the original partition.
- mapPartitionsWithIndex(Function2<Object, Iterator<T>, Iterator<U>>, boolean, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- mapPartitionsWithIndex$default$2() - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- mapPartitionsWithIndex$default$2() - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- mapPartitionsWithIndex$default$2() - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- mapPartitionsWithIndex$default$2() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- mapPartitionsWithIndex$default$2() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- mapPartitionsWithIndex$default$2() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- mapPartitionsWithIndex$default$2() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- mapPartitionsWithIndex$default$2() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- mapPartitionsWithIndex$default$2() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- mapPartitionsWithIndex$default$2() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- mapPartitionsWithIndex$default$2() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- mapPartitionsWithIndex$default$2() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- mapPartitionsWithIndex$default$2() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- mapPartitionsWithIndexInternal$default$2() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- mapPartitionsWithIndexInternal$default$2() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- mapPartitionsWithIndexInternal$default$2() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- mapPartitionsWithIndexInternal$default$2() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- mapPartitionsWithIndexInternal$default$2() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- mapPartitionsWithIndexInternal$default$2() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- mapPartitionsWithIndexInternal$default$2() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- mapPartitionsWithIndexInternal$default$2() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- mapPartitionsWithIndexInternal$default$2() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- mapPartitionsWithIndexInternal$default$2() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- mapPartitionsWithInputSplit(Function2<InputSplit, Iterator<Tuple2<K, V>>, Iterator<R>>, boolean) - 类 中的方法org.apache.spark.api.java.JavaHadoopRDD
-
Maps over a partition, providing the InputSplit that was used as the base of the partition.
- mapPartitionsWithInputSplit(Function2<InputSplit, Iterator<Tuple2<K, V>>, Iterator<R>>, boolean) - 类 中的方法org.apache.spark.api.java.JavaNewHadoopRDD
-
Maps over a partition, providing the InputSplit that was used as the base of the partition.
- mapPartitionsWithInputSplit(Function2<InputSplit, Iterator<Tuple2<K, V>>, Iterator<U>>, boolean, ClassTag<U>) - 类 中的方法org.apache.spark.rdd.HadoopRDD
-
Maps over a partition, providing the InputSplit that was used as the base of the partition.
- mapPartitionsWithInputSplit(Function2<InputSplit, Iterator<Tuple2<K, V>>, Iterator<U>>, boolean, ClassTag<U>) - 类 中的方法org.apache.spark.rdd.NewHadoopRDD
-
Maps over a partition, providing the InputSplit that was used as the base of the partition.
- mapredInputFormat() - 类 中的方法org.apache.spark.scheduler.InputFormatInfo
-
- mapreduceInputFormat() - 类 中的方法org.apache.spark.scheduler.InputFormatInfo
-
- mapSideCombine() - 类 中的方法org.apache.spark.ShuffleDependency
-
- mapStatuses() - 类 中的方法org.apache.spark.ShuffleStatus
-
MapStatus for each partition.
- mapToDouble(DoubleFunction<T>) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- mapToDouble(DoubleFunction<T>) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- mapToDouble(DoubleFunction<T>) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- mapToDouble(DoubleFunction<T>) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Return a new RDD by applying a function to all elements of this RDD.
- mapToJson(Map<String, String>) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
------------------------------ *
Util JSON serialization methods |
- mapToPair(PairFunction<T, K2, V2>) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- mapToPair(PairFunction<T, K2, V2>) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- mapToPair(PairFunction<T, K2, V2>) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- mapToPair(PairFunction<T, K2, V2>) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Return a new RDD by applying a function to all elements of this RDD.
- mapToPair(PairFunction<T, K2, V2>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- mapToPair(PairFunction<T, K2, V2>) - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return a new DStream by applying a function to all elements of this DStream.
- mapToPair(PairFunction<T, K2, V2>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- mapToPair(PairFunction<T, K2, V2>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- mapToPair(PairFunction<T, K2, V2>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- mapToPair(PairFunction<T, K2, V2>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- mapToPair(PairFunction<T, K2, V2>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- mapTriplets(Function1<EdgeTriplet<VD, ED>, ED2>, ClassTag<ED2>) - 类 中的方法org.apache.spark.graphx.Graph
-
Transforms each edge attribute using the map function, passing it the adjacent vertex
attributes as well.
- mapTriplets(Function1<EdgeTriplet<VD, ED>, ED2>, TripletFields, ClassTag<ED2>) - 类 中的方法org.apache.spark.graphx.Graph
-
Transforms each edge attribute using the map function, passing it the adjacent vertex
attributes as well.
- mapTriplets(Function2<Object, Iterator<EdgeTriplet<VD, ED>>, Iterator<ED2>>, TripletFields, ClassTag<ED2>) - 类 中的方法org.apache.spark.graphx.Graph
-
Transforms each edge attribute a partition at a time using the map function, passing it the
adjacent vertex attributes as well.
- mapTriplets(Function2<Object, Iterator<EdgeTriplet<VD, ED>>, Iterator<ED2>>, TripletFields, ClassTag<ED2>) - 类 中的方法org.apache.spark.graphx.impl.GraphImpl
-
- MapType - org.apache.spark.sql.types中的类
-
The data type for Maps.
- MapType(DataType, DataType, boolean) - 类 的构造器org.apache.spark.sql.types.MapType
-
- MapType() - 类 的构造器org.apache.spark.sql.types.MapType
-
No-arg constructor for kryo.
- mapValues(Function<V, U>) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Pass each value in the key-value pair RDD through a map function without changing the keys;
this also retains the original RDD's partitioning.
- mapValues(Function1<Edge<ED>, ED2>, ClassTag<ED2>) - 类 中的方法org.apache.spark.graphx.EdgeRDD
-
Map the values in an edge partitioning preserving the structure but changing the values.
- mapValues(Function1<Edge<ED>, ED2>, ClassTag<ED2>) - 类 中的方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- mapValues(Function1<VD, VD2>, ClassTag<VD2>) - 类 中的方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- mapValues(Function2<Object, VD, VD2>, ClassTag<VD2>) - 类 中的方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- mapValues(Function1<VD, VD2>, ClassTag<VD2>) - 类 中的方法org.apache.spark.graphx.VertexRDD
-
Maps each vertex attribute, preserving the index.
- mapValues(Function2<Object, VD, VD2>, ClassTag<VD2>) - 类 中的方法org.apache.spark.graphx.VertexRDD
-
Maps each vertex attribute, additionally supplying the vertex ID.
- mapValues(Function1<V, U>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Pass each value in the key-value pair RDD through a map function without changing the keys;
this also retains the original RDD's partitioning.
- mapValues(Function1<V, W>, Encoder<W>) - 类 中的方法org.apache.spark.sql.KeyValueGroupedDataset
-
- mapValues(MapFunction<V, W>, Encoder<W>) - 类 中的方法org.apache.spark.sql.KeyValueGroupedDataset
-
- mapValues(Function<V, U>) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying a map function to the value of each key-value pairs in
'this' DStream without changing the key.
- mapValues(Function<V, U>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- mapValues(Function<V, U>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- mapValues(Function1<V, U>, ClassTag<U>) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying a map function to the value of each key-value pairs in
'this' DStream without changing the key.
- mapVertices(Function2<Object, VD, VD2>, ClassTag<VD2>, Predef.$eq$colon$eq<VD, VD2>) - 类 中的方法org.apache.spark.graphx.Graph
-
Transforms each vertex attribute in the graph using the map function.
- mapVertices(Function2<Object, VD, VD2>, ClassTag<VD2>, Predef.$eq$colon$eq<VD, VD2>) - 类 中的方法org.apache.spark.graphx.impl.GraphImpl
-
- mapVertices$default$3(Function2<Object, VD, VD2>) - 类 中的静态方法org.apache.spark.graphx.impl.GraphImpl
-
- mapWithState(StateSpec<K, V, StateType, MappedType>) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
:: Experimental ::
Return a
JavaMapWithStateDStream by applying a function to every key-value element of
this stream, while maintaining some state data for each unique key.
- mapWithState(StateSpec<K, V, StateType, MappedType>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- mapWithState(StateSpec<K, V, StateType, MappedType>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- mapWithState(StateSpec<K, V, StateType, MappedType>, ClassTag<StateType>, ClassTag<MappedType>) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
:: Experimental ::
Return a
MapWithStateDStream by applying a function to every key-value element of
this stream, while maintaining some state data for each unique key.
- MapWithStateDStream<KeyType,ValueType,StateType,MappedType> - org.apache.spark.streaming.dstream中的类
-
:: Experimental ::
DStream representing the stream of data generated by
mapWithState operation on a
pair DStream.
- MapWithStateDStream(StreamingContext, ClassTag<MappedType>) - 类 的构造器org.apache.spark.streaming.dstream.MapWithStateDStream
-
- mark(int) - 类 中的方法org.apache.spark.storage.BufferReleasingInputStream
-
- markSupported() - 类 中的方法org.apache.spark.storage.BufferReleasingInputStream
-
- mask(Graph<VD2, ED2>, ClassTag<VD2>, ClassTag<ED2>) - 类 中的方法org.apache.spark.graphx.Graph
-
Restricts the graph to only the vertices and edges that are also in other, but keeps the
attributes from this graph.
- mask(Graph<VD2, ED2>, ClassTag<VD2>, ClassTag<ED2>) - 类 中的方法org.apache.spark.graphx.impl.GraphImpl
-
- master() - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
- master() - 类 中的方法org.apache.spark.SparkContext
-
- master(String) - 类 中的方法org.apache.spark.sql.SparkSession.Builder
-
Sets the Spark master URL to connect to, such as "local" to run locally, "local[4]" to
run locally with 4 cores, or "spark://master:7077" to run on a Spark standalone cluster.
- Matrices - org.apache.spark.ml.linalg中的类
-
- Matrices() - 类 的构造器org.apache.spark.ml.linalg.Matrices
-
- Matrices - org.apache.spark.mllib.linalg中的类
-
- Matrices() - 类 的构造器org.apache.spark.mllib.linalg.Matrices
-
- Matrix - org.apache.spark.ml.linalg中的接口
-
Trait for a local matrix.
- Matrix - org.apache.spark.mllib.linalg中的接口
-
Trait for a local matrix.
- MatrixEntry - org.apache.spark.mllib.linalg.distributed中的类
-
Represents an entry in a distributed matrix.
- MatrixEntry(long, long, double) - 类 的构造器org.apache.spark.mllib.linalg.distributed.MatrixEntry
-
- MatrixFactorizationModel - org.apache.spark.mllib.recommendation中的类
-
Model representing the result of matrix factorization.
- MatrixFactorizationModel(int, RDD<Tuple2<Object, double[]>>, RDD<Tuple2<Object, double[]>>) - 类 的构造器org.apache.spark.mllib.recommendation.MatrixFactorizationModel
-
- MatrixFactorizationModel.SaveLoadV1_0$ - org.apache.spark.mllib.recommendation中的类
-
- MatrixImplicits - org.apache.spark.mllib.linalg中的类
-
Implicit methods available in Scala for converting
Matrix to
Matrix and vice versa.
- MatrixImplicits() - 类 的构造器org.apache.spark.mllib.linalg.MatrixImplicits
-
- MatrixType() - 类 中的静态方法org.apache.spark.ml.linalg.SQLDataTypes
-
- max() - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
Returns the maximum element from this RDD as defined by
the default comparator natural order.
- max(Comparator<T>) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- max(Comparator<T>) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- max(Comparator<T>) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Returns the maximum element from this RDD as defined by the specified
Comparator[T].
- max(Ordering<T>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- max(Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- max(Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- max(Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- max(Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- MAX() - 类 中的静态方法org.apache.spark.ml.attribute.AttributeKeys
-
- max() - 类 中的方法org.apache.spark.ml.attribute.NumericAttribute
-
- max() - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScaler
-
- max() - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScalerModel
-
- max(Column, Column) - 类 中的静态方法org.apache.spark.ml.stat.Summarizer
-
- max(Column) - 类 中的静态方法org.apache.spark.ml.stat.Summarizer
-
- max() - 类 中的方法org.apache.spark.mllib.stat.MultivariateOnlineSummarizer
-
Maximum value of each dimension.
- max() - 接口 中的方法org.apache.spark.mllib.stat.MultivariateStatisticalSummary
-
Maximum value of each column.
- max(Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- max(Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- max(Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- max(Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- max(Ordering<T>) - 类 中的方法org.apache.spark.rdd.RDD
-
Returns the max of this RDD as defined by the implicit Ordering[T].
- max(Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- max(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the maximum value of the expression in a group.
- max(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the maximum value of the column in a group.
- max(String...) - 类 中的方法org.apache.spark.sql.RelationalGroupedDataset
-
Compute the max value for each numeric columns for each group.
- max(Seq<String>) - 类 中的方法org.apache.spark.sql.RelationalGroupedDataset
-
Compute the max value for each numeric columns for each group.
- max(Ordering<B>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- max(Duration) - 类 中的方法org.apache.spark.streaming.Duration
-
- max(Time) - 类 中的方法org.apache.spark.streaming.Time
-
- max(long, long) - 类 中的静态方法org.apache.spark.streaming.util.RawTextHelper
-
- max(long) - 类 中的方法org.apache.spark.util.kvstore.KVStoreView
-
Stops iteration after a number of elements has been retrieved.
- max() - 类 中的方法org.apache.spark.util.StatCounter
-
- MAX_FEATURES_FOR_NORMAL_SOLVER() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
When using LinearRegression.solver == "normal", the solver must limit the number of
features to at most this number.
- MAX_INT_DIGITS() - 类 中的静态方法org.apache.spark.sql.types.Decimal
-
Maximum number of decimal digits an Int can represent
- MAX_LONG_DIGITS() - 类 中的静态方法org.apache.spark.sql.types.Decimal
-
Maximum number of decimal digits a Long can represent
- MAX_PRECISION() - 类 中的静态方法org.apache.spark.sql.types.DecimalType
-
- MAX_RETAINED_DEAD_EXECUTORS() - 类 中的静态方法org.apache.spark.status.config
-
- MAX_RETAINED_JOBS() - 类 中的静态方法org.apache.spark.status.config
-
- MAX_RETAINED_ROOT_NODES() - 类 中的静态方法org.apache.spark.status.config
-
- MAX_RETAINED_STAGES() - 类 中的静态方法org.apache.spark.status.config
-
- MAX_RETAINED_TASKS_PER_STAGE() - 类 中的静态方法org.apache.spark.status.config
-
- MAX_SCALE() - 类 中的静态方法org.apache.spark.sql.types.DecimalType
-
- maxAbs() - 类 中的方法org.apache.spark.ml.feature.MaxAbsScalerModel
-
- MaxAbsScaler - org.apache.spark.ml.feature中的类
-
Rescale each feature individually to range [-1, 1] by dividing through the largest maximum
absolute value in each feature.
- MaxAbsScaler(String) - 类 的构造器org.apache.spark.ml.feature.MaxAbsScaler
-
- MaxAbsScaler() - 类 的构造器org.apache.spark.ml.feature.MaxAbsScaler
-
- MaxAbsScalerModel - org.apache.spark.ml.feature中的类
-
- maxBins() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- maxBins() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- maxBins() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- maxBins() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- maxBins() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- maxBins() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- maxBins() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- maxBins() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- maxBins() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- maxBins() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- maxBins() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- maxBins() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- maxBins() - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- maxBufferSizeMb() - 类 中的方法org.apache.spark.serializer.KryoSerializer
-
- maxBy(Function1<A, B>, Ordering<B>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- maxCategories() - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexer
-
- maxCategories() - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexerModel
-
- maxCores() - 类 中的方法org.apache.spark.status.api.v1.ApplicationInfo
-
- maxDepth() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- maxDepth() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- maxDepth() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- maxDepth() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- maxDepth() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- maxDepth() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- maxDepth() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- maxDepth() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- maxDepth() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- maxDepth() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- maxDepth() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- maxDepth() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- maxDepth() - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- maxId() - 类 中的静态方法org.apache.spark.mllib.tree.configuration.Algo
-
- maxId() - 类 中的静态方法org.apache.spark.mllib.tree.configuration.EnsembleCombiningStrategy
-
- maxId() - 类 中的静态方法org.apache.spark.mllib.tree.configuration.FeatureType
-
- maxId() - 类 中的静态方法org.apache.spark.mllib.tree.configuration.QuantileStrategy
-
- maxId() - 类 中的静态方法org.apache.spark.rdd.CheckpointState
-
- maxId() - 类 中的静态方法org.apache.spark.scheduler.SchedulingMode
-
- maxId() - 类 中的静态方法org.apache.spark.scheduler.TaskLocality
-
- maxId() - 类 中的静态方法org.apache.spark.streaming.scheduler.ReceiverState
-
- maxId() - 类 中的静态方法org.apache.spark.TaskState
-
- maxIter() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- maxIter() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- maxIter() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- maxIter() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- maxIter() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- maxIter() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- maxIter() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- maxIter() - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeans
-
- maxIter() - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
- maxIter() - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- maxIter() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixture
-
- maxIter() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- maxIter() - 类 中的静态方法org.apache.spark.ml.clustering.KMeans
-
- maxIter() - 类 中的静态方法org.apache.spark.ml.clustering.KMeansModel
-
- maxIter() - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- maxIter() - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- maxIter() - 类 中的静态方法org.apache.spark.ml.feature.Word2Vec
-
- maxIter() - 类 中的静态方法org.apache.spark.ml.feature.Word2VecModel
-
- maxIter() - 接口 中的方法org.apache.spark.ml.param.shared.HasMaxIter
-
Param for maximum number of iterations (>= 0).
- maxIter() - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- maxIter() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- maxIter() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- maxIter() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- maxIter() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- maxIter() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- maxIter() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- maxIter() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- maxIter() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- maxIters() - 类 中的方法org.apache.spark.graphx.lib.SVDPlusPlus.Conf
-
- maxMem() - 类 中的方法org.apache.spark.scheduler.SparkListenerBlockManagerAdded
-
- maxMem() - 类 中的方法org.apache.spark.storage.StorageStatus
-
已过时。
Return the max memory can be used by this block manager.
- maxMemory() - 类 中的方法org.apache.spark.status.api.v1.ExecutorSummary
-
- maxMemory() - 类 中的方法org.apache.spark.status.LiveExecutor
-
- maxMemory() - 类 中的方法org.apache.spark.storage.StorageStatus
-
已过时。
- maxMemoryInMB() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- maxMemoryInMB() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- maxMemoryInMB() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- maxMemoryInMB() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- maxMemoryInMB() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- maxMemoryInMB() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- maxMemoryInMB() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- maxMemoryInMB() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- maxMemoryInMB() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- maxMemoryInMB() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- maxMemoryInMB() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- maxMemoryInMB() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- maxMemoryInMB() - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- maxMessageSizeBytes(SparkConf) - 类 中的静态方法org.apache.spark.util.RpcUtils
-
Returns the configured max message size for messages in bytes.
- maxNodesInLevel(int) - 类 中的静态方法org.apache.spark.mllib.tree.model.Node
-
Return the maximum number of nodes which can be in the given level of the tree.
- maxOffHeapMem() - 类 中的方法org.apache.spark.scheduler.SparkListenerBlockManagerAdded
-
- maxOffHeapMem() - 类 中的方法org.apache.spark.storage.StorageStatus
-
已过时。
- maxOffHeapMemSize() - 类 中的方法org.apache.spark.storage.BlockManagerMessages.RegisterBlockManager
-
- maxOnHeapMem() - 类 中的方法org.apache.spark.scheduler.SparkListenerBlockManagerAdded
-
- maxOnHeapMem() - 类 中的方法org.apache.spark.storage.StorageStatus
-
已过时。
- maxOnHeapMemSize() - 类 中的方法org.apache.spark.storage.BlockManagerMessages.RegisterBlockManager
-
- maxReplicas() - 类 中的方法org.apache.spark.storage.BlockManagerMessages.ReplicateBlock
-
- maxRows() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- maxRows() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- maxRows() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- maxRowsPerPartition() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- maxRowsPerPartition() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- maxRowsPerPartition() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- maxSentenceLength() - 类 中的静态方法org.apache.spark.ml.feature.Word2Vec
-
- maxSentenceLength() - 类 中的静态方法org.apache.spark.ml.feature.Word2VecModel
-
- maxTasks() - 类 中的方法org.apache.spark.status.api.v1.ExecutorSummary
-
- maxTasks() - 类 中的方法org.apache.spark.status.LiveExecutor
-
- maxVal() - 类 中的方法org.apache.spark.graphx.lib.SVDPlusPlus.Conf
-
- maybeUpdateOutputMetrics(OutputMetrics, Function0<Object>, long) - 类 中的静态方法org.apache.spark.internal.io.SparkHadoopWriterUtils
-
- md5(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Calculates the MD5 digest of a binary column and returns the value
as a 32 character hex string.
- mean() - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
Compute the mean of this RDD's elements.
- mean() - 类 中的方法org.apache.spark.ml.feature.StandardScalerModel
-
- mean() - 类 中的方法org.apache.spark.ml.stat.distribution.MultivariateGaussian
-
- mean(Column, Column) - 类 中的静态方法org.apache.spark.ml.stat.Summarizer
-
- mean(Column) - 类 中的静态方法org.apache.spark.ml.stat.Summarizer
-
- mean() - 类 中的方法org.apache.spark.mllib.feature.StandardScalerModel
-
- mean() - 类 中的方法org.apache.spark.mllib.random.ExponentialGenerator
-
- mean() - 类 中的方法org.apache.spark.mllib.random.LogNormalGenerator
-
- mean() - 类 中的方法org.apache.spark.mllib.random.PoissonGenerator
-
- mean() - 类 中的方法org.apache.spark.mllib.stat.MultivariateOnlineSummarizer
-
Sample mean of each dimension.
- mean() - 接口 中的方法org.apache.spark.mllib.stat.MultivariateStatisticalSummary
-
Sample mean vector.
- mean() - 类 中的方法org.apache.spark.partial.BoundedDouble
-
- mean() - 类 中的方法org.apache.spark.rdd.DoubleRDDFunctions
-
Compute the mean of this RDD's elements.
- mean(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the average of the values in a group.
- mean(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the average of the values in a group.
- mean(String...) - 类 中的方法org.apache.spark.sql.RelationalGroupedDataset
-
Compute the average value for each numeric columns for each group.
- mean(Seq<String>) - 类 中的方法org.apache.spark.sql.RelationalGroupedDataset
-
Compute the average value for each numeric columns for each group.
- mean() - 类 中的方法org.apache.spark.util.StatCounter
-
- meanAbsoluteError() - 类 中的方法org.apache.spark.ml.regression.LinearRegressionSummary
-
Returns the mean absolute error, which is a risk function corresponding to the
expected value of the absolute error loss or l1-norm loss.
- meanAbsoluteError() - 类 中的方法org.apache.spark.mllib.evaluation.RegressionMetrics
-
Returns the mean absolute error, which is a risk function corresponding to the
expected value of the absolute error loss or l1-norm loss.
- meanApprox(long, Double) - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
Return the approximate mean of the elements in this RDD.
- meanApprox(long) - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
Approximate operation to return the mean within a timeout.
- meanApprox(long, double) - 类 中的方法org.apache.spark.rdd.DoubleRDDFunctions
-
Approximate operation to return the mean within a timeout.
- meanAveragePrecision() - 类 中的方法org.apache.spark.mllib.evaluation.RankingMetrics
-
Returns the mean average precision (MAP) of all the queries.
- means() - 类 中的方法org.apache.spark.ml.clustering.ExpectationAggregator
-
- means() - 类 中的方法org.apache.spark.mllib.clustering.ExpectationSum
-
- meanSquaredError() - 类 中的方法org.apache.spark.ml.regression.LinearRegressionSummary
-
Returns the mean squared error, which is a risk function corresponding to the
expected value of the squared error loss or quadratic loss.
- meanSquaredError() - 类 中的方法org.apache.spark.mllib.evaluation.RegressionMetrics
-
Returns the mean squared error, which is a risk function corresponding to the
expected value of the squared error loss or quadratic loss.
- megabytesToString(long) - 类 中的静态方法org.apache.spark.util.Utils
-
Convert a quantity in megabytes to a human-readable string such as "4.0 MB".
- MEM_SPILL() - 类 中的静态方法org.apache.spark.status.TaskIndexNames
-
- MEMORY_AND_DISK - 类 中的静态变量org.apache.spark.api.java.StorageLevels
-
- MEMORY_AND_DISK() - 类 中的静态方法org.apache.spark.storage.StorageLevel
-
- MEMORY_AND_DISK_2 - 类 中的静态变量org.apache.spark.api.java.StorageLevels
-
- MEMORY_AND_DISK_2() - 类 中的静态方法org.apache.spark.storage.StorageLevel
-
- MEMORY_AND_DISK_SER - 类 中的静态变量org.apache.spark.api.java.StorageLevels
-
- MEMORY_AND_DISK_SER() - 类 中的静态方法org.apache.spark.storage.StorageLevel
-
- MEMORY_AND_DISK_SER_2 - 类 中的静态变量org.apache.spark.api.java.StorageLevels
-
- MEMORY_AND_DISK_SER_2() - 类 中的静态方法org.apache.spark.storage.StorageLevel
-
- MEMORY_BYTES_SPILLED() - 类 中的静态方法org.apache.spark.InternalAccumulator
-
- MEMORY_ONLY - 类 中的静态变量org.apache.spark.api.java.StorageLevels
-
- MEMORY_ONLY() - 类 中的静态方法org.apache.spark.storage.StorageLevel
-
- MEMORY_ONLY_2 - 类 中的静态变量org.apache.spark.api.java.StorageLevels
-
- MEMORY_ONLY_2() - 类 中的静态方法org.apache.spark.storage.StorageLevel
-
- MEMORY_ONLY_SER - 类 中的静态变量org.apache.spark.api.java.StorageLevels
-
- MEMORY_ONLY_SER() - 类 中的静态方法org.apache.spark.storage.StorageLevel
-
- MEMORY_ONLY_SER_2 - 类 中的静态变量org.apache.spark.api.java.StorageLevels
-
- MEMORY_ONLY_SER_2() - 类 中的静态方法org.apache.spark.storage.StorageLevel
-
- memoryBytesSpilled() - 类 中的方法org.apache.spark.status.api.v1.ExecutorStageSummary
-
- memoryBytesSpilled() - 类 中的方法org.apache.spark.status.api.v1.StageData
-
- memoryBytesSpilled() - 类 中的方法org.apache.spark.status.api.v1.TaskMetricDistributions
-
- memoryBytesSpilled() - 类 中的方法org.apache.spark.status.api.v1.TaskMetrics
-
- memoryCost(int, int) - 类 中的静态方法org.apache.spark.mllib.feature.PCAUtil
-
- MemoryEntry<T> - org.apache.spark.storage.memory中的接口
-
- memoryManager() - 类 中的方法org.apache.spark.SparkEnv
-
- memoryMetrics() - 类 中的方法org.apache.spark.status.api.v1.ExecutorSummary
-
- MemoryMetrics - org.apache.spark.status.api.v1中的类
-
- memoryMode() - 类 中的方法org.apache.spark.storage.memory.DeserializedMemoryEntry
-
- memoryMode() - 接口 中的方法org.apache.spark.storage.memory.MemoryEntry
-
- memoryMode() - 类 中的方法org.apache.spark.storage.memory.SerializedMemoryEntry
-
- MemoryParam - org.apache.spark.util中的类
-
An extractor object for parsing JVM memory strings, such as "10g", into an Int representing
the number of megabytes.
- MemoryParam() - 类 的构造器org.apache.spark.util.MemoryParam
-
- memoryPerExecutorMB() - 类 中的方法org.apache.spark.status.api.v1.ApplicationInfo
-
- memoryRemaining() - 类 中的方法org.apache.spark.status.api.v1.RDDDataDistribution
-
- memoryStringToMb(String) - 类 中的静态方法org.apache.spark.util.Utils
-
Convert a Java memory parameter passed to -Xmx (such as 300m or 1g) to a number of mebibytes.
- memoryUsed() - 类 中的方法org.apache.spark.status.api.v1.ExecutorSummary
-
- memoryUsed() - 类 中的方法org.apache.spark.status.api.v1.RDDDataDistribution
-
- memoryUsed() - 类 中的方法org.apache.spark.status.api.v1.RDDPartitionInfo
-
- memoryUsed() - 类 中的方法org.apache.spark.status.api.v1.RDDStorageInfo
-
- memoryUsed() - 类 中的方法org.apache.spark.status.LiveExecutor
-
- memoryUsed() - 类 中的方法org.apache.spark.status.LiveRDD
-
- memoryUsed() - 类 中的方法org.apache.spark.status.LiveRDDDistribution
-
- memoryUsed() - 类 中的方法org.apache.spark.status.LiveRDDPartition
-
- memoryUsedBytes() - 类 中的方法org.apache.spark.sql.streaming.StateOperatorProgress
-
- memRemaining() - 类 中的方法org.apache.spark.storage.StorageStatus
-
已过时。
Return the memory remaining in this block manager.
- memSize() - 类 中的方法org.apache.spark.storage.BlockManagerMessages.UpdateBlockInfo
-
- memSize() - 类 中的方法org.apache.spark.storage.BlockStatus
-
- memSize() - 类 中的方法org.apache.spark.storage.BlockUpdatedInfo
-
- memSize() - 类 中的方法org.apache.spark.storage.RDDInfo
-
- memUsed() - 类 中的方法org.apache.spark.storage.StorageStatus
-
已过时。
Return the memory used by this block manager.
- memUsedByRdd(int) - 类 中的方法org.apache.spark.storage.StorageStatus
-
已过时。
Return the memory used by the given RDD in this block manager in O(1) time.
- merge(R) - 类 中的方法org.apache.spark.Accumulable
-
已过时。
Merge two accumulable objects together
Normally, a user will not want to use this version, but will instead call add.
- merge(R) - 类 中的静态方法org.apache.spark.Accumulator
-
已过时。
- merge(ExpectationAggregator) - 类 中的方法org.apache.spark.ml.clustering.ExpectationAggregator
-
Merge another ExpectationAggregator, update the weights, means and covariances
for each distributions, and update the log likelihood.
- merge(AFTAggregator) - 类 中的方法org.apache.spark.ml.regression.AFTAggregator
-
Merge another AFTAggregator, and update the loss and gradient
of the objective function.
- merge(IDF.DocumentFrequencyAggregator) - 类 中的方法org.apache.spark.mllib.feature.IDF.DocumentFrequencyAggregator
-
Merges another.
- merge(MultivariateOnlineSummarizer) - 类 中的方法org.apache.spark.mllib.stat.MultivariateOnlineSummarizer
-
Merge another MultivariateOnlineSummarizer, and update the statistical summary.
- merge(BUF, BUF) - 类 中的方法org.apache.spark.sql.expressions.Aggregator
-
Merge two intermediate values.
- merge(MutableAggregationBuffer, Row) - 类 中的方法org.apache.spark.sql.expressions.UserDefinedAggregateFunction
-
Merges two aggregation buffers and stores the updated buffer values back to buffer1.
- merge(AccumulatorV2<IN, OUT>) - 类 中的方法org.apache.spark.util.AccumulatorV2
-
Merges another same-type accumulator into this one and update its state, i.e. this should be
merge-in-place.
- merge(AccumulatorV2<T, List<T>>) - 类 中的方法org.apache.spark.util.CollectionAccumulator
-
- merge(AccumulatorV2<Double, Double>) - 类 中的方法org.apache.spark.util.DoubleAccumulator
-
- merge(AccumulatorV2<T, R>) - 类 中的方法org.apache.spark.util.LegacyAccumulatorWrapper
-
- merge(AccumulatorV2<Long, Long>) - 类 中的方法org.apache.spark.util.LongAccumulator
-
- merge(double) - 类 中的方法org.apache.spark.util.StatCounter
-
Add a value into this StatCounter, updating the internal statistics.
- merge(TraversableOnce<Object>) - 类 中的方法org.apache.spark.util.StatCounter
-
Add multiple values into this StatCounter, updating the internal statistics.
- merge(StatCounter) - 类 中的方法org.apache.spark.util.StatCounter
-
Merge another StatCounter into this one, adding up the internal statistics.
- mergeCombiners() - 类 中的方法org.apache.spark.Aggregator
-
- mergeInPlace(BloomFilter) - 类 中的方法org.apache.spark.util.sketch.BloomFilter
-
Combines this bloom filter with another bloom filter by performing a bitwise OR of the
underlying data.
- mergeInPlace(CountMinSketch) - 类 中的方法org.apache.spark.util.sketch.CountMinSketch
-
- mergeOffsets(PartitionOffset[]) - 接口 中的方法org.apache.spark.sql.sources.v2.reader.streaming.ContinuousReader
-
Merge partitioned offsets coming from
ContinuousDataReader instances for each
partition to a single global offset.
- mergeValue() - 类 中的方法org.apache.spark.Aggregator
-
- message() - 类 中的方法org.apache.spark.FetchFailed
-
- message() - 类 中的方法org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RegisterExecutorFailed
-
- message() - 类 中的方法org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RemoveWorker
-
- message() - 类 中的静态方法org.apache.spark.scheduler.ExecutorKilled
-
- message() - 类 中的静态方法org.apache.spark.scheduler.LossReasonPending
-
- message() - 异常错误 中的方法org.apache.spark.sql.AnalysisException
-
- message() - 异常错误 中的方法org.apache.spark.sql.streaming.StreamingQueryException
-
- message() - 类 中的方法org.apache.spark.sql.streaming.StreamingQueryStatus
-
- MetaAlgorithmReadWrite - org.apache.spark.ml.util中的类
-
Default Meta-Algorithm read and write implementation.
- MetaAlgorithmReadWrite() - 类 的构造器org.apache.spark.ml.util.MetaAlgorithmReadWrite
-
- Metadata - org.apache.spark.sql.types中的类
-
Metadata is a wrapper over Map[String, Any] that limits the value type to simple ones: Boolean,
Long, Double, String, Metadata, Array[Boolean], Array[Long], Array[Double], Array[String], and
Array[Metadata].
- metadata() - 类 中的方法org.apache.spark.sql.types.StructField
-
- metadata() - 类 中的方法org.apache.spark.streaming.scheduler.StreamInputInfo
-
- METADATA_KEY_DESCRIPTION() - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamInputInfo
-
The key for description in StreamInputInfo.metadata.
- MetadataBuilder - org.apache.spark.sql.types中的类
-
- MetadataBuilder() - 类 的构造器org.apache.spark.sql.types.MetadataBuilder
-
- metadataDescription() - 类 中的方法org.apache.spark.streaming.scheduler.StreamInputInfo
-
- MetadataUtils - org.apache.spark.ml.util中的类
-
Helper utilities for algorithms using ML metadata
- MetadataUtils() - 类 的构造器org.apache.spark.ml.util.MetadataUtils
-
- Method(String, Function2<Object, Object, Object>) - 类 的构造器org.apache.spark.mllib.stat.test.ChiSqTest.Method
-
- method() - 类 中的方法org.apache.spark.mllib.stat.test.ChiSqTestResult
-
- Method$() - 类 的构造器org.apache.spark.mllib.stat.test.ChiSqTest.Method$
-
- MethodIdentifier<T> - org.apache.spark.util中的类
-
Helper class to identify a method.
- MethodIdentifier(Class<T>, String, String) - 类 的构造器org.apache.spark.util.MethodIdentifier
-
- methodName() - 类 中的静态方法org.apache.spark.mllib.stat.test.StudentTTest
-
- methodName() - 类 中的静态方法org.apache.spark.mllib.stat.test.WelchTTest
-
- METRIC_COMPILATION_TIME() - 类 中的静态方法org.apache.spark.metrics.source.CodegenMetrics
-
Histogram of the time it took to compile source code text (in milliseconds).
- METRIC_FILE_CACHE_HITS() - 类 中的静态方法org.apache.spark.metrics.source.HiveCatalogMetrics
-
Tracks the total number of files served from the file status cache instead of discovered.
- METRIC_FILES_DISCOVERED() - 类 中的静态方法org.apache.spark.metrics.source.HiveCatalogMetrics
-
Tracks the total number of files discovered off of the filesystem by InMemoryFileIndex.
- METRIC_GENERATED_CLASS_BYTECODE_SIZE() - 类 中的静态方法org.apache.spark.metrics.source.CodegenMetrics
-
Histogram of the bytecode size of each class generated by CodeGenerator.
- METRIC_GENERATED_METHOD_BYTECODE_SIZE() - 类 中的静态方法org.apache.spark.metrics.source.CodegenMetrics
-
Histogram of the bytecode size of each method in classes generated by CodeGenerator.
- METRIC_HIVE_CLIENT_CALLS() - 类 中的静态方法org.apache.spark.metrics.source.HiveCatalogMetrics
-
Tracks the total number of Hive client calls (e.g. to lookup a table).
- METRIC_PARALLEL_LISTING_JOB_COUNT() - 类 中的静态方法org.apache.spark.metrics.source.HiveCatalogMetrics
-
Tracks the total number of Spark jobs launched for parallel file listing.
- METRIC_PARTITIONS_FETCHED() - 类 中的静态方法org.apache.spark.metrics.source.HiveCatalogMetrics
-
Tracks the total number of partition metadata entries fetched via the client api.
- METRIC_SOURCE_CODE_SIZE() - 类 中的静态方法org.apache.spark.metrics.source.CodegenMetrics
-
Histogram of the length of source code text compiled by CodeGenerator (in characters).
- metricName() - 类 中的方法org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
-
param for metric name in evaluation (supports "areaUnderROC" (default), "areaUnderPR")
- metricName() - 类 中的方法org.apache.spark.ml.evaluation.ClusteringEvaluator
-
param for metric name in evaluation
(supports "silhouette" (default))
- metricName() - 类 中的方法org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
param for metric name in evaluation (supports "f1" (default), "weightedPrecision",
"weightedRecall", "accuracy")
- metricName() - 类 中的方法org.apache.spark.ml.evaluation.RegressionEvaluator
-
Param for metric name in evaluation.
- metricRegistry() - 类 中的静态方法org.apache.spark.metrics.source.CodegenMetrics
-
- metricRegistry() - 类 中的静态方法org.apache.spark.metrics.source.HiveCatalogMetrics
-
- metrics(String...) - 类 中的静态方法org.apache.spark.ml.stat.Summarizer
-
Given a list of metrics, provides a builder that it turns computes metrics from a column.
- metrics(Seq<String>) - 类 中的静态方法org.apache.spark.ml.stat.Summarizer
-
Given a list of metrics, provides a builder that it turns computes metrics from a column.
- metrics() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- metrics() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- metrics() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- metrics() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- metrics() - 类 中的方法org.apache.spark.status.LiveExecutorStageSummary
-
- metrics() - 类 中的方法org.apache.spark.status.LiveStage
-
- METRICS_PREFIX() - 类 中的静态方法org.apache.spark.InternalAccumulator
-
- metricsSystem() - 类 中的方法org.apache.spark.SparkEnv
-
- MFDataGenerator - org.apache.spark.mllib.util中的类
-
:: DeveloperApi ::
Generate RDD(s) containing data for Matrix Factorization.
- MFDataGenerator() - 类 的构造器org.apache.spark.mllib.util.MFDataGenerator
-
- MicroBatchReader - org.apache.spark.sql.sources.v2.reader.streaming中的接口
-
- MicroBatchReadSupport - org.apache.spark.sql.sources.v2中的接口
-
- microF1Measure() - 类 中的方法org.apache.spark.mllib.evaluation.MultilabelMetrics
-
Returns micro-averaged label-based f1-measure
(equals to micro-averaged document-based f1-measure)
- microPrecision() - 类 中的方法org.apache.spark.mllib.evaluation.MultilabelMetrics
-
Returns micro-averaged label-based precision
(equals to micro-averaged document-based precision)
- microRecall() - 类 中的方法org.apache.spark.mllib.evaluation.MultilabelMetrics
-
Returns micro-averaged label-based recall
(equals to micro-averaged document-based recall)
- mightContain(Object) - 类 中的方法org.apache.spark.util.sketch.BloomFilter
-
Returns true if the element might have been put in this Bloom filter,
false if this is definitely not the case.
- mightContainBinary(byte[]) - 类 中的方法org.apache.spark.util.sketch.BloomFilter
-
- mightContainLong(long) - 类 中的方法org.apache.spark.util.sketch.BloomFilter
-
- mightContainString(String) - 类 中的方法org.apache.spark.util.sketch.BloomFilter
-
- milliseconds() - 类 中的方法org.apache.spark.streaming.Duration
-
- milliseconds(long) - 类 中的静态方法org.apache.spark.streaming.Durations
-
- Milliseconds - org.apache.spark.streaming中的类
-
Helper object that creates instance of
Duration representing
a given number of milliseconds.
- Milliseconds() - 类 的构造器org.apache.spark.streaming.Milliseconds
-
- milliseconds() - 类 中的方法org.apache.spark.streaming.Time
-
- millisToString(long) - 类 中的静态方法org.apache.spark.scheduler.StatsReportListener
-
Reformat a time interval in milliseconds to a prettier format for output
- min() - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
Returns the minimum element from this RDD as defined by
the default comparator natural order.
- min(Comparator<T>) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- min(Comparator<T>) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- min(Comparator<T>) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Returns the minimum element from this RDD as defined by the specified
Comparator[T].
- min(Ordering<T>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- min(Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- min(Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- min(Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- min(Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- MIN() - 类 中的静态方法org.apache.spark.ml.attribute.AttributeKeys
-
- min() - 类 中的方法org.apache.spark.ml.attribute.NumericAttribute
-
- min() - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScaler
-
- min() - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScalerModel
-
- min(Column, Column) - 类 中的静态方法org.apache.spark.ml.stat.Summarizer
-
- min(Column) - 类 中的静态方法org.apache.spark.ml.stat.Summarizer
-
- min() - 类 中的方法org.apache.spark.mllib.stat.MultivariateOnlineSummarizer
-
Minimum value of each dimension.
- min() - 接口 中的方法org.apache.spark.mllib.stat.MultivariateStatisticalSummary
-
Minimum value of each column.
- min(Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- min(Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- min(Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- min(Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- min(Ordering<T>) - 类 中的方法org.apache.spark.rdd.RDD
-
Returns the min of this RDD as defined by the implicit Ordering[T].
- min(Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- min(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the minimum value of the expression in a group.
- min(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the minimum value of the column in a group.
- min(String...) - 类 中的方法org.apache.spark.sql.RelationalGroupedDataset
-
Compute the min value for each numeric column for each group.
- min(Seq<String>) - 类 中的方法org.apache.spark.sql.RelationalGroupedDataset
-
Compute the min value for each numeric column for each group.
- min(Ordering<B>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- min(Duration) - 类 中的方法org.apache.spark.streaming.Duration
-
- min(Time) - 类 中的方法org.apache.spark.streaming.Time
-
- min() - 类 中的方法org.apache.spark.util.StatCounter
-
- minBy(Function1<A, B>, Ordering<B>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- minConfidence() - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowth
-
- minConfidence() - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowthModel
-
- minCount() - 类 中的静态方法org.apache.spark.ml.feature.Word2Vec
-
- minCount() - 类 中的静态方法org.apache.spark.ml.feature.Word2VecModel
-
- minDF() - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizer
-
- minDF() - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizerModel
-
- minDivisibleClusterSize() - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeans
-
- minDivisibleClusterSize() - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
- minDocFreq() - 类 中的静态方法org.apache.spark.ml.feature.IDF
-
- minDocFreq() - 类 中的静态方法org.apache.spark.ml.feature.IDFModel
-
- minDocFreq() - 类 中的方法org.apache.spark.mllib.feature.IDF.DocumentFrequencyAggregator
-
- minDocFreq() - 类 中的方法org.apache.spark.mllib.feature.IDF
-
- MinHashLSH - org.apache.spark.ml.feature中的类
-
:: Experimental ::
LSH class for Jaccard distance.
- MinHashLSH(String) - 类 的构造器org.apache.spark.ml.feature.MinHashLSH
-
- MinHashLSH() - 类 的构造器org.apache.spark.ml.feature.MinHashLSH
-
- MinHashLSHModel - org.apache.spark.ml.feature中的类
-
:: Experimental ::
Model produced by
MinHashLSH, where multiple hash functions are stored.
- MINIMUM_ADJUSTED_SCALE() - 类 中的静态方法org.apache.spark.sql.types.DecimalType
-
- minInfoGain() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- minInfoGain() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- minInfoGain() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- minInfoGain() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- minInfoGain() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- minInfoGain() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- minInfoGain() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- minInfoGain() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- minInfoGain() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- minInfoGain() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- minInfoGain() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- minInfoGain() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- minInfoGain() - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- minInstancesPerNode() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- minInstancesPerNode() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- minInstancesPerNode() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- minInstancesPerNode() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- minInstancesPerNode() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- minInstancesPerNode() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- minInstancesPerNode() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- minInstancesPerNode() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- minInstancesPerNode() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- minInstancesPerNode() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- minInstancesPerNode() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- minInstancesPerNode() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- minInstancesPerNode() - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- MinMax() - 类 中的静态方法org.apache.spark.mllib.tree.configuration.QuantileStrategy
-
- MinMaxScaler - org.apache.spark.ml.feature中的类
-
Rescale each feature individually to a common range [min, max] linearly using column summary
statistics, which is also known as min-max normalization or Rescaling.
- MinMaxScaler(String) - 类 的构造器org.apache.spark.ml.feature.MinMaxScaler
-
- MinMaxScaler() - 类 的构造器org.apache.spark.ml.feature.MinMaxScaler
-
- MinMaxScalerModel - org.apache.spark.ml.feature中的类
-
- minorVersion(String) - 类 中的静态方法org.apache.spark.util.VersionUtils
-
Given a Spark version string, return the minor version number.
- minSamplingRate() - 类 中的静态方法org.apache.spark.util.random.BinomialBounds
-
- minSupport() - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowth
-
- minSupport() - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowthModel
-
- minTF() - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizer
-
- minTF() - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizerModel
-
- minTokenLength() - 类 中的方法org.apache.spark.ml.feature.RegexTokenizer
-
Minimum token length, greater than or equal to 0.
- minus(RDD<Tuple2<Object, VD>>) - 类 中的方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- minus(VertexRDD<VD>) - 类 中的方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- minus(RDD<Tuple2<Object, VD>>) - 类 中的方法org.apache.spark.graphx.VertexRDD
-
For each VertexId present in both this and other, minus will act as a set difference
operation returning only those unique VertexId's present in this.
- minus(VertexRDD<VD>) - 类 中的方法org.apache.spark.graphx.VertexRDD
-
For each VertexId present in both this and other, minus will act as a set difference
operation returning only those unique VertexId's present in this.
- minus(Object) - 类 中的方法org.apache.spark.sql.Column
-
Subtraction.
- minus(Duration) - 类 中的方法org.apache.spark.streaming.Duration
-
- minus(Time) - 类 中的方法org.apache.spark.streaming.Time
-
- minus(Duration) - 类 中的方法org.apache.spark.streaming.Time
-
- minute(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Extracts the minutes as an integer from a given date/timestamp/string.
- minutes() - 类 中的静态方法org.apache.spark.scheduler.StatsReportListener
-
- minutes(long) - 类 中的静态方法org.apache.spark.streaming.Durations
-
- Minutes - org.apache.spark.streaming中的类
-
Helper object that creates instance of
Duration representing
a given number of minutes.
- Minutes() - 类 的构造器org.apache.spark.streaming.Minutes
-
- minVal() - 类 中的方法org.apache.spark.graphx.lib.SVDPlusPlus.Conf
-
- missingInput() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- missingInput() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- missingInput() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- missingInput() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- missingValue() - 类 中的静态方法org.apache.spark.ml.feature.Imputer
-
- missingValue() - 类 中的静态方法org.apache.spark.ml.feature.ImputerModel
-
- mkList() - 类 中的静态方法org.apache.spark.ml.feature.RFormulaParser
-
- mkString() - 接口 中的方法org.apache.spark.sql.Row
-
Displays all elements of this sequence in a string (without a separator).
- mkString(String) - 接口 中的方法org.apache.spark.sql.Row
-
Displays all elements of this sequence in a string using a separator string.
- mkString(String, String, String) - 接口 中的方法org.apache.spark.sql.Row
-
Displays all elements of this traversable or iterator in a string using
start, end, and separator strings.
- mkString(String, String, String) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- mkString(String) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- mkString() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- ML_ATTR() - 类 中的静态方法org.apache.spark.ml.attribute.AttributeKeys
-
- mlDenseMatrixToMLlibDenseMatrix(DenseMatrix) - 类 中的静态方法org.apache.spark.mllib.linalg.MatrixImplicits
-
- mlDenseVectorToMLlibDenseVector(DenseVector) - 类 中的静态方法org.apache.spark.mllib.linalg.VectorImplicits
-
- mllibDenseMatrixToMLDenseMatrix(DenseMatrix) - 类 中的静态方法org.apache.spark.mllib.linalg.MatrixImplicits
-
- mllibDenseVectorToMLDenseVector(DenseVector) - 类 中的静态方法org.apache.spark.mllib.linalg.VectorImplicits
-
- mllibMatrixToMLMatrix(Matrix) - 类 中的静态方法org.apache.spark.mllib.linalg.MatrixImplicits
-
- mllibSparseMatrixToMLSparseMatrix(SparseMatrix) - 类 中的静态方法org.apache.spark.mllib.linalg.MatrixImplicits
-
- mllibSparseVectorToMLSparseVector(SparseVector) - 类 中的静态方法org.apache.spark.mllib.linalg.VectorImplicits
-
- mllibVectorToMLVector(Vector) - 类 中的静态方法org.apache.spark.mllib.linalg.VectorImplicits
-
- mlMatrixToMLlibMatrix(Matrix) - 类 中的静态方法org.apache.spark.mllib.linalg.MatrixImplicits
-
- MLPairRDDFunctions<K,V> - org.apache.spark.mllib.rdd中的类
-
:: DeveloperApi ::
Machine learning specific Pair RDD functions.
- MLPairRDDFunctions(RDD<Tuple2<K, V>>, ClassTag<K>, ClassTag<V>) - 类 的构造器org.apache.spark.mllib.rdd.MLPairRDDFunctions
-
- MLReadable<T> - org.apache.spark.ml.util中的接口
-
Trait for objects that provide MLReader.
- MLReader<T> - org.apache.spark.ml.util中的类
-
Abstract class for utility classes that can load ML instances.
- MLReader() - 类 的构造器org.apache.spark.ml.util.MLReader
-
- mlSparseMatrixToMLlibSparseMatrix(SparseMatrix) - 类 中的静态方法org.apache.spark.mllib.linalg.MatrixImplicits
-
- mlSparseVectorToMLlibSparseVector(SparseVector) - 类 中的静态方法org.apache.spark.mllib.linalg.VectorImplicits
-
- MLUtils - org.apache.spark.mllib.util中的类
-
Helper methods to load, save and pre-process data used in MLLib.
- MLUtils() - 类 的构造器org.apache.spark.mllib.util.MLUtils
-
- mlVectorToMLlibVector(Vector) - 类 中的静态方法org.apache.spark.mllib.linalg.VectorImplicits
-
- MLWritable - org.apache.spark.ml.util中的接口
-
Trait for classes that provide MLWriter.
- MLWriter - org.apache.spark.ml.util中的类
-
Abstract class for utility classes that can save ML instances.
- MLWriter() - 类 的构造器org.apache.spark.ml.util.MLWriter
-
- mod(Object) - 类 中的方法org.apache.spark.sql.Column
-
Modulo (a.k.a. remainder) expression.
- mode(SaveMode) - 类 中的方法org.apache.spark.sql.DataFrameWriter
-
Specifies the behavior when data or table already exists.
- mode(String) - 类 中的方法org.apache.spark.sql.DataFrameWriter
-
Specifies the behavior when data or table already exists.
- mode() - 类 中的方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- Model<M extends Model<M>> - org.apache.spark.ml中的类
-
- Model() - 类 的构造器org.apache.spark.ml.Model
-
- models() - 类 中的方法org.apache.spark.ml.classification.OneVsRestModel
-
- modelType() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- modelType() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- modelType() - 类 中的方法org.apache.spark.mllib.classification.NaiveBayesModel
-
- modelType() - 类 中的方法org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV2_0$.Data
-
- MODULE$ - 类 中的静态变量org.apache.spark.AccumulatorParam.DoubleAccumulatorParam$
-
已过时。
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.AccumulatorParam.FloatAccumulatorParam$
-
已过时。
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.AccumulatorParam.IntAccumulatorParam$
-
已过时。
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.AccumulatorParam.LongAccumulatorParam$
-
已过时。
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.AccumulatorParam.StringAccumulatorParam$
-
已过时。
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.graphx.PartitionStrategy.CanonicalRandomVertexCut$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.graphx.PartitionStrategy.EdgePartition1D$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.graphx.PartitionStrategy.EdgePartition2D$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.graphx.PartitionStrategy.RandomVertexCut$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.internal.io.FileCommitProtocol.EmptyTaskCommitMessage$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.InternalAccumulator.input$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.InternalAccumulator.output$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.InternalAccumulator.shuffleRead$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.InternalAccumulator.shuffleWrite$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.ml.evaluation.SquaredEuclideanSilhouette.ClusterStats$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.ml.feature.Word2VecModel.Word2VecModelWriter$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.ml.Pipeline.SharedReadWrite$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.ml.recommendation.ALS.InBlock$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.ml.recommendation.ALS.Rating$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.ml.recommendation.ALS.RatingBlock$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.ml.regression.GeneralizedLinearRegression.Binomial$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.ml.regression.GeneralizedLinearRegression.CLogLog$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.ml.regression.GeneralizedLinearRegression.Family$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.ml.regression.GeneralizedLinearRegression.FamilyAndLink$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.ml.regression.GeneralizedLinearRegression.Gamma$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.ml.regression.GeneralizedLinearRegression.Gaussian$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.ml.regression.GeneralizedLinearRegression.Identity$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.ml.regression.GeneralizedLinearRegression.Inverse$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.ml.regression.GeneralizedLinearRegression.Link$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.ml.regression.GeneralizedLinearRegression.Log$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.ml.regression.GeneralizedLinearRegression.Logit$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.ml.regression.GeneralizedLinearRegression.Poisson$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.ml.regression.GeneralizedLinearRegression.Probit$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.ml.regression.GeneralizedLinearRegression.Sqrt$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.ml.regression.GeneralizedLinearRegression.Tweedie$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.ml.tree.DecisionTreeModelReadWrite.NodeData$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.ml.tree.DecisionTreeModelReadWrite.SplitData$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.ml.tree.EnsembleModelReadWrite.EnsembleNodeData$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.mllib.classification.impl.GLMClassificationModel.SaveLoadV1_0$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV1_0$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV2_0$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.mllib.clustering.BisectingKMeansModel.SaveLoadV1_0$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.mllib.clustering.KMeansModel.SaveLoadV1_0$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.mllib.clustering.PowerIterationClustering.Assignment$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.mllib.clustering.PowerIterationClusteringModel.SaveLoadV1_0$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.mllib.feature.ChiSqSelectorModel.SaveLoadV1_0$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.mllib.fpm.FPGrowthModel.SaveLoadV1_0$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.mllib.fpm.PrefixSpan.Postfix$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.mllib.fpm.PrefixSpan.Prefix$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.mllib.fpm.PrefixSpanModel.SaveLoadV1_0$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.mllib.recommendation.MatrixFactorizationModel.SaveLoadV1_0$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.mllib.regression.impl.GLMRegressionModel.SaveLoadV1_0$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.mllib.stat.test.ChiSqTest.Method$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.mllib.stat.test.ChiSqTest.NullHypothesis$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.mllib.stat.test.KolmogorovSmirnovTest.NullHypothesis$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.rdd.HadoopRDD.HadoopMapPartitionsWithSplitRDD$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.rdd.NewHadoopRDD.NewHadoopMapPartitionsWithSplitRDD$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.AddWebUIFilter$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.GetExecutorLossReason$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.KillExecutors$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.KillExecutorsOnHost$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.KillTask$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.LaunchTask$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RegisterClusterManager$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RegisteredExecutor$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RegisterExecutor$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RegisterExecutorFailed$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RemoveExecutor$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RemoveWorker$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RequestExecutors$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RetrieveLastAllocatedExecutorId$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RetrieveSparkAppConfig$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.ReviveOffers$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.SetupDriver$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.Shutdown$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.SparkAppConfig$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.StatusUpdate$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.StopDriver$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.StopExecutor$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.StopExecutors$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.UpdateDelegationTokens$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.serializer.SerializationDebugger.ObjectStreamClassMethods$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.sql.hive.HiveShim.HiveFunctionWrapper$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.sql.RelationalGroupedDataset.CubeType$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.sql.RelationalGroupedDataset.GroupByType$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.sql.RelationalGroupedDataset.PivotType$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.sql.RelationalGroupedDataset.RollupType$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.sql.types.Decimal.DecimalAsIfIntegral$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.sql.types.Decimal.DecimalIsFractional$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.sql.types.DecimalType.Expression$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.sql.types.DecimalType.Fixed$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.storage.BlockManagerMessages.BlockLocationsAndStatus$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.storage.BlockManagerMessages.BlockManagerHeartbeat$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.storage.BlockManagerMessages.GetBlockStatus$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.storage.BlockManagerMessages.GetExecutorEndpointRef$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.storage.BlockManagerMessages.GetLocations$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.storage.BlockManagerMessages.GetLocationsAndStatus$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.storage.BlockManagerMessages.GetLocationsMultipleBlockIds$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.storage.BlockManagerMessages.GetMatchingBlockIds$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.storage.BlockManagerMessages.GetMemoryStatus$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.storage.BlockManagerMessages.GetPeers$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.storage.BlockManagerMessages.GetStorageStatus$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.storage.BlockManagerMessages.HasCachedBlocks$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.storage.BlockManagerMessages.RegisterBlockManager$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.storage.BlockManagerMessages.RemoveBlock$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.storage.BlockManagerMessages.RemoveBroadcast$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.storage.BlockManagerMessages.RemoveExecutor$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.storage.BlockManagerMessages.RemoveRdd$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.storage.BlockManagerMessages.RemoveShuffle$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.storage.BlockManagerMessages.ReplicateBlock$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.storage.BlockManagerMessages.StopBlockManagerMaster$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.storage.BlockManagerMessages.TriggerThreadDump$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.storage.BlockManagerMessages.UpdateBlockInfo$
-
Static reference to the singleton instance of this Scala object.
- MODULE$ - 类 中的静态变量org.apache.spark.ui.JettyUtils.ServletParams$
-
Static reference to the singleton instance of this Scala object.
- monotonically_increasing_id() - 类 中的静态方法org.apache.spark.sql.functions
-
A column expression that generates monotonically increasing 64-bit integers.
- monotonicallyIncreasingId() - 类 中的静态方法org.apache.spark.sql.functions
-
- month(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Extracts the month as an integer from a given date/timestamp/string.
- months_between(Column, Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns number of months between dates date1 and date2.
- msDurationToString(long) - 类 中的静态方法org.apache.spark.util.Utils
-
Returns a human-readable string representing a duration such as "35ms"
- MsSqlServerDialect - org.apache.spark.sql.jdbc中的类
-
- MsSqlServerDialect() - 类 的构造器org.apache.spark.sql.jdbc.MsSqlServerDialect
-
- mu() - 类 中的方法org.apache.spark.mllib.stat.distribution.MultivariateGaussian
-
- MulticlassClassificationEvaluator - org.apache.spark.ml.evaluation中的类
-
:: Experimental ::
Evaluator for multiclass classification, which expects two input columns: prediction and label.
- MulticlassClassificationEvaluator(String) - 类 的构造器org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- MulticlassClassificationEvaluator() - 类 的构造器org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- multiclassMetrics() - 接口 中的方法org.apache.spark.ml.classification.LogisticRegressionSummary
-
- MulticlassMetrics - org.apache.spark.mllib.evaluation中的类
-
Evaluator for multiclass classification.
- MulticlassMetrics(RDD<Tuple2<Object, Object>>) - 类 的构造器org.apache.spark.mllib.evaluation.MulticlassMetrics
-
- MultilabelMetrics - org.apache.spark.mllib.evaluation中的类
-
Evaluator for multilabel classification.
- MultilabelMetrics(RDD<Tuple2<double[], double[]>>) - 类 的构造器org.apache.spark.mllib.evaluation.MultilabelMetrics
-
- multiLabelValidator(int) - 类 中的静态方法org.apache.spark.mllib.util.DataValidators
-
Function to check if labels used for k class multi-label classification are
in the range of {0, 1, ..., k - 1}.
- MultilayerPerceptronClassificationModel - org.apache.spark.ml.classification中的类
-
Classification model based on the Multilayer Perceptron.
- MultilayerPerceptronClassifier - org.apache.spark.ml.classification中的类
-
Classifier trainer based on the Multilayer Perceptron.
- MultilayerPerceptronClassifier(String) - 类 的构造器org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- MultilayerPerceptronClassifier() - 类 的构造器org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- multiply(DenseMatrix) - 类 中的静态方法org.apache.spark.ml.linalg.DenseMatrix
-
- multiply(DenseVector) - 类 中的静态方法org.apache.spark.ml.linalg.DenseMatrix
-
- multiply(Vector) - 类 中的静态方法org.apache.spark.ml.linalg.DenseMatrix
-
- multiply(DenseMatrix) - 接口 中的方法org.apache.spark.ml.linalg.Matrix
-
Convenience method for Matrix-DenseMatrix multiplication.
- multiply(DenseVector) - 接口 中的方法org.apache.spark.ml.linalg.Matrix
-
Convenience method for Matrix-DenseVector multiplication.
- multiply(Vector) - 接口 中的方法org.apache.spark.ml.linalg.Matrix
-
Convenience method for Matrix-Vector multiplication.
- multiply(DenseMatrix) - 类 中的静态方法org.apache.spark.ml.linalg.SparseMatrix
-
- multiply(DenseVector) - 类 中的静态方法org.apache.spark.ml.linalg.SparseMatrix
-
- multiply(Vector) - 类 中的静态方法org.apache.spark.ml.linalg.SparseMatrix
-
- multiply(DenseMatrix) - 类 中的静态方法org.apache.spark.mllib.linalg.DenseMatrix
-
- multiply(DenseVector) - 类 中的静态方法org.apache.spark.mllib.linalg.DenseMatrix
-
- multiply(Vector) - 类 中的静态方法org.apache.spark.mllib.linalg.DenseMatrix
-
- multiply(BlockMatrix) - 类 中的方法org.apache.spark.mllib.linalg.distributed.BlockMatrix
-
- multiply(BlockMatrix, int) - 类 中的方法org.apache.spark.mllib.linalg.distributed.BlockMatrix
-
- multiply(Matrix) - 类 中的方法org.apache.spark.mllib.linalg.distributed.IndexedRowMatrix
-
Multiply this matrix by a local matrix on the right.
- multiply(Matrix) - 类 中的方法org.apache.spark.mllib.linalg.distributed.RowMatrix
-
Multiply this matrix by a local matrix on the right.
- multiply(DenseMatrix) - 接口 中的方法org.apache.spark.mllib.linalg.Matrix
-
Convenience method for Matrix-DenseMatrix multiplication.
- multiply(DenseVector) - 接口 中的方法org.apache.spark.mllib.linalg.Matrix
-
Convenience method for Matrix-DenseVector multiplication.
- multiply(Vector) - 接口 中的方法org.apache.spark.mllib.linalg.Matrix
-
Convenience method for Matrix-Vector multiplication.
- multiply(DenseMatrix) - 类 中的静态方法org.apache.spark.mllib.linalg.SparseMatrix
-
- multiply(DenseVector) - 类 中的静态方法org.apache.spark.mllib.linalg.SparseMatrix
-
- multiply(Vector) - 类 中的静态方法org.apache.spark.mllib.linalg.SparseMatrix
-
- multiply(Object) - 类 中的方法org.apache.spark.sql.Column
-
Multiplication of this expression and another expression.
- MultivariateGaussian - org.apache.spark.ml.stat.distribution中的类
-
This class provides basic functionality for a Multivariate Gaussian (Normal) Distribution.
- MultivariateGaussian(Vector, Matrix) - 类 的构造器org.apache.spark.ml.stat.distribution.MultivariateGaussian
-
- MultivariateGaussian - org.apache.spark.mllib.stat.distribution中的类
-
:: DeveloperApi ::
This class provides basic functionality for a Multivariate Gaussian (Normal) Distribution.
- MultivariateGaussian(Vector, Matrix) - 类 的构造器org.apache.spark.mllib.stat.distribution.MultivariateGaussian
-
- MultivariateOnlineSummarizer - org.apache.spark.mllib.stat中的类
-
:: DeveloperApi ::
MultivariateOnlineSummarizer implements
MultivariateStatisticalSummary to compute the mean,
variance, minimum, maximum, counts, and nonzero counts for instances in sparse or dense vector
format in an online fashion.
- MultivariateOnlineSummarizer() - 类 的构造器org.apache.spark.mllib.stat.MultivariateOnlineSummarizer
-
- MultivariateStatisticalSummary - org.apache.spark.mllib.stat中的接口
-
Trait for multivariate statistical summary of a data matrix.
- MutableAggregationBuffer - org.apache.spark.sql.expressions中的类
-
A Row representing a mutable aggregation buffer.
- MutableAggregationBuffer() - 类 的构造器org.apache.spark.sql.expressions.MutableAggregationBuffer
-
- MutablePair<T1,T2> - org.apache.spark.util中的类
-
:: DeveloperApi ::
A tuple of 2 elements.
- MutablePair(T1, T2) - 类 的构造器org.apache.spark.util.MutablePair
-
- MutablePair() - 类 的构造器org.apache.spark.util.MutablePair
-
No-arg constructor for serialization
- myName() - 类 中的方法org.apache.spark.util.InnerClosureFinder
-
- MySQLDialect - org.apache.spark.sql.jdbc中的类
-
- MySQLDialect() - 类 的构造器org.apache.spark.sql.jdbc.MySQLDialect
-
- p() - 类 中的方法org.apache.spark.ml.feature.Normalizer
-
Normalization in L^p^ space.
- p(int) - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- p(int) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- p(int) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- p(int) - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- padTo(int, B, CanBuildFrom<Repr, B, That>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- pageRank(double, double) - 类 中的方法org.apache.spark.graphx.GraphOps
-
Run a dynamic version of PageRank returning a graph with vertex attributes containing the
PageRank and edge attributes containing the normalized edge weight.
- PageRank - org.apache.spark.graphx.lib中的类
-
PageRank algorithm implementation.
- PageRank() - 类 的构造器org.apache.spark.graphx.lib.PageRank
-
- PairDStreamFunctions<K,V> - org.apache.spark.streaming.dstream中的类
-
Extra functions available on DStream of (key, value) pairs through an implicit conversion.
- PairDStreamFunctions(DStream<Tuple2<K, V>>, ClassTag<K>, ClassTag<V>, Ordering<K>) - 类 的构造器org.apache.spark.streaming.dstream.PairDStreamFunctions
-
- PairFlatMapFunction<T,K,V> - org.apache.spark.api.java.function中的接口
-
A function that returns zero or more key-value pair records from each input record.
- PairFunction<T,K,V> - org.apache.spark.api.java.function中的接口
-
A function that returns key-value pairs (Tuple2<K, V>), and can be used to
construct PairRDDs.
- PairRDDFunctions<K,V> - org.apache.spark.rdd中的类
-
Extra functions available on RDDs of (key, value) pairs through an implicit conversion.
- PairRDDFunctions(RDD<Tuple2<K, V>>, ClassTag<K>, ClassTag<V>, Ordering<K>) - 类 的构造器org.apache.spark.rdd.PairRDDFunctions
-
- PairwiseRRDD<T> - org.apache.spark.api.r中的类
-
Form an RDD[(Int, Array[Byte])] from key-value pairs returned from R.
- PairwiseRRDD(RDD<T>, int, byte[], String, byte[], Object[], ClassTag<T>) - 类 的构造器org.apache.spark.api.r.PairwiseRRDD
-
- par() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- parallelism() - 类 中的静态方法org.apache.spark.ml.classification.OneVsRest
-
- parallelism() - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidator
-
- parallelism() - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplit
-
- parallelize(List<T>, int) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
Distribute a local Scala collection to form an RDD.
- parallelize(List<T>) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
Distribute a local Scala collection to form an RDD.
- parallelize(Seq<T>, int, ClassTag<T>) - 类 中的方法org.apache.spark.SparkContext
-
Distribute a local Scala collection to form an RDD.
- parallelizeDoubles(List<Double>, int) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
Distribute a local Scala collection to form an RDD.
- parallelizeDoubles(List<Double>) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
Distribute a local Scala collection to form an RDD.
- parallelizePairs(List<Tuple2<K, V>>, int) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
Distribute a local Scala collection to form an RDD.
- parallelizePairs(List<Tuple2<K, V>>) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
Distribute a local Scala collection to form an RDD.
- Param<T> - org.apache.spark.ml.param中的类
-
:: DeveloperApi ::
A param with self-contained documentation and optionally default value.
- Param(String, String, String, Function1<T, Object>) - 类 的构造器org.apache.spark.ml.param.Param
-
- Param(Identifiable, String, String, Function1<T, Object>) - 类 的构造器org.apache.spark.ml.param.Param
-
- Param(String, String, String) - 类 的构造器org.apache.spark.ml.param.Param
-
- Param(Identifiable, String, String) - 类 的构造器org.apache.spark.ml.param.Param
-
- param() - 类 中的方法org.apache.spark.ml.param.ParamPair
-
- ParamGridBuilder - org.apache.spark.ml.tuning中的类
-
Builder for a param grid used in grid search-based model selection.
- ParamGridBuilder() - 类 的构造器org.apache.spark.ml.tuning.ParamGridBuilder
-
- ParamMap - org.apache.spark.ml.param中的类
-
A param to value map.
- ParamMap() - 类 的构造器org.apache.spark.ml.param.ParamMap
-
Creates an empty param map.
- paramMap() - 接口 中的方法org.apache.spark.ml.param.Params
-
Internal param map for user-supplied values.
- ParamPair<T> - org.apache.spark.ml.param中的类
-
A param and its value.
- ParamPair(Param<T>, T) - 类 的构造器org.apache.spark.ml.param.ParamPair
-
- params() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- params() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- params() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- params() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- params() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- params() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- params() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- params() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- params() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- params() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- params() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- params() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- params() - 类 中的静态方法org.apache.spark.ml.classification.OneVsRest
-
- params() - 类 中的静态方法org.apache.spark.ml.classification.OneVsRestModel
-
- params() - 类 中的静态方法org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- params() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- params() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- params() - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeans
-
- params() - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
- params() - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- params() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixture
-
- params() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- params() - 类 中的静态方法org.apache.spark.ml.clustering.KMeans
-
- params() - 类 中的静态方法org.apache.spark.ml.clustering.KMeansModel
-
- params() - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- params() - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- params() - 类 中的静态方法org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
-
- params() - 类 中的静态方法org.apache.spark.ml.evaluation.ClusteringEvaluator
-
- params() - 类 中的静态方法org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- params() - 类 中的静态方法org.apache.spark.ml.evaluation.RegressionEvaluator
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.Binarizer
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.Bucketizer
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelector
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.ColumnPruner
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizer
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizerModel
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.DCT
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.ElementwiseProduct
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.FeatureHasher
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.HashingTF
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.IDF
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.IDFModel
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.Imputer
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.ImputerModel
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.IndexToString
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.Interaction
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.MaxAbsScaler
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.MaxAbsScalerModel
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSH
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSHModel
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScaler
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScalerModel
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.NGram
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.Normalizer
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoder
-
已过时。
- params() - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderEstimator
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderModel
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.PCA
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.PCAModel
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.PolynomialExpansion
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.QuantileDiscretizer
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.RegexTokenizer
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.RFormula
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.RFormulaModel
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.SQLTransformer
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.StandardScaler
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.StandardScalerModel
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.StopWordsRemover
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.StringIndexer
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.StringIndexerModel
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.Tokenizer
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.VectorAssembler
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.VectorAttributeRewriter
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexer
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexerModel
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.VectorSizeHint
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.VectorSlicer
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.Word2Vec
-
- params() - 类 中的静态方法org.apache.spark.ml.feature.Word2VecModel
-
- params() - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowth
-
- params() - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowthModel
-
- Params - org.apache.spark.ml.param中的接口
-
:: DeveloperApi ::
Trait for components that take parameters.
- params() - 接口 中的方法org.apache.spark.ml.param.Params
-
Returns all params sorted by their names.
- params() - 类 中的静态方法org.apache.spark.ml.Pipeline
-
- params() - 类 中的静态方法org.apache.spark.ml.PipelineModel
-
- params() - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- params() - 类 中的静态方法org.apache.spark.ml.recommendation.ALSModel
-
- params() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- params() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- params() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- params() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- params() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- params() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- params() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- params() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- params() - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegression
-
- params() - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegressionModel
-
- params() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- params() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- params() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- params() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- params() - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidator
-
- params() - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidatorModel
-
- params() - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplit
-
- params() - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplitModel
-
- ParamValidators - org.apache.spark.ml.param中的类
-
:: DeveloperApi ::
Factory methods for common validation functions for Param.isValid.
- ParamValidators() - 类 的构造器org.apache.spark.ml.param.ParamValidators
-
- parent() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.classification.OneVsRestModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.clustering.KMeansModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.feature.Bucketizer
-
- parent() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizerModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.feature.IDFModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.feature.ImputerModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.feature.MaxAbsScalerModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSHModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScalerModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.feature.PCAModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.feature.RFormulaModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.feature.StandardScalerModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.feature.StringIndexerModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexerModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.feature.Word2VecModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowthModel
-
- parent() - 类 中的方法org.apache.spark.ml.Model
-
The parent estimator that produced this model.
- parent() - 类 中的静态方法org.apache.spark.ml.param.DoubleParam
-
- parent() - 类 中的静态方法org.apache.spark.ml.param.FloatParam
-
- parent() - 类 中的方法org.apache.spark.ml.param.Param
-
- parent() - 类 中的静态方法org.apache.spark.ml.PipelineModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.recommendation.ALSModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegressionModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidatorModel
-
- parent() - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplitModel
-
- parent(Object) - 类 中的方法org.apache.spark.util.kvstore.KVStoreView
-
Defines the value of the parent index when iterating over a child index.
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.classification.OneVsRestModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.clustering.KMeansModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.Bucketizer
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizerModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.IDFModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.ImputerModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.MaxAbsScalerModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSHModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScalerModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.PCAModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.StandardScalerModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.StringIndexerModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexerModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.Word2VecModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowthModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.PipelineModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.recommendation.ALSModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegressionModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidatorModel
-
- parent_$eq(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplitModel
-
- parentIds() - 类 中的方法org.apache.spark.scheduler.StageInfo
-
- parentIds() - 类 中的方法org.apache.spark.storage.RDDInfo
-
- parentIndex(int) - 类 中的静态方法org.apache.spark.mllib.tree.model.Node
-
Get the parent index of the given node, or 0 if it is the root.
- parentState() - 类 中的静态方法org.apache.spark.sql.hive.HiveSessionStateBuilder
-
- parquet(String...) - 类 中的方法org.apache.spark.sql.DataFrameReader
-
Loads a Parquet file, returning the result as a DataFrame.
- parquet(String) - 类 中的方法org.apache.spark.sql.DataFrameReader
-
Loads a Parquet file, returning the result as a DataFrame.
- parquet(Seq<String>) - 类 中的方法org.apache.spark.sql.DataFrameReader
-
Loads a Parquet file, returning the result as a DataFrame.
- parquet(String) - 类 中的方法org.apache.spark.sql.DataFrameWriter
-
Saves the content of the DataFrame in Parquet format at the specified path.
- parquet(String) - 类 中的方法org.apache.spark.sql.streaming.DataStreamReader
-
Loads a Parquet file stream, returning the result as a DataFrame.
- parquetFile(String...) - 类 中的方法org.apache.spark.sql.SQLContext
-
- parquetFile(Seq<String>) - 类 中的方法org.apache.spark.sql.SQLContext
-
- parse(String) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaParser
-
- parse(String) - 类 中的静态方法org.apache.spark.mllib.linalg.Vectors
-
Parses a string resulted from
Vector.toString into a
Vector.
- parse(String) - 类 中的静态方法org.apache.spark.mllib.regression.LabeledPoint
-
Parses a string resulted from
LabeledPoint#toString into
an
LabeledPoint.
- parse(String) - 类 中的静态方法org.apache.spark.mllib.util.NumericParser
-
Parses a string into a Double, an Array[Double], or a Seq[Any].
- parseAll(Parsers.Parser<T>, Reader<Object>) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaParser
-
- parseAll(Parsers.Parser<T>, Reader) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaParser
-
- parseAll(Parsers.Parser<T>, CharSequence) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaParser
-
- parseHostPort(String) - 类 中的静态方法org.apache.spark.util.Utils
-
- parseIgnoreCase(Class<E>, String) - 类 中的静态方法org.apache.spark.util.EnumUtil
-
- Parser(Function1<Reader<Object>, Parsers.ParseResult<T>>) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaParser
-
- parseStandaloneMasterUrls(String) - 类 中的静态方法org.apache.spark.util.Utils
-
Split the comma delimited string of master URLs into a list.
- PartialResult<R> - org.apache.spark.partial中的类
-
- PartialResult(R, boolean) - 类 的构造器org.apache.spark.partial.PartialResult
-
- Partition - org.apache.spark中的接口
-
An identifier for a partition in an RDD.
- partition() - 类 中的方法org.apache.spark.scheduler.AskPermissionToCommitOutput
-
- partition() - 类 中的方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- partition(Function1<A, Object>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- partition(String) - 类 中的方法org.apache.spark.status.LiveRDD
-
- partitionBy(Partitioner) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Return a copy of the RDD partitioned using the specified partitioner.
- partitionBy(PartitionStrategy) - 类 中的方法org.apache.spark.graphx.Graph
-
Repartitions the edges in the graph according to partitionStrategy.
- partitionBy(PartitionStrategy, int) - 类 中的方法org.apache.spark.graphx.Graph
-
Repartitions the edges in the graph according to partitionStrategy.
- partitionBy(PartitionStrategy) - 类 中的方法org.apache.spark.graphx.impl.GraphImpl
-
- partitionBy(PartitionStrategy, int) - 类 中的方法org.apache.spark.graphx.impl.GraphImpl
-
- partitionBy(Partitioner) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Return a copy of the RDD partitioned using the specified partitioner.
- partitionBy(String...) - 类 中的方法org.apache.spark.sql.DataFrameWriter
-
Partitions the output by the given columns on the file system.
- partitionBy(Seq<String>) - 类 中的方法org.apache.spark.sql.DataFrameWriter
-
Partitions the output by the given columns on the file system.
- partitionBy(String, String...) - 类 中的静态方法org.apache.spark.sql.expressions.Window
-
Creates a
WindowSpec with the partitioning defined.
- partitionBy(Column...) - 类 中的静态方法org.apache.spark.sql.expressions.Window
-
Creates a
WindowSpec with the partitioning defined.
- partitionBy(String, Seq<String>) - 类 中的静态方法org.apache.spark.sql.expressions.Window
-
Creates a
WindowSpec with the partitioning defined.
- partitionBy(Seq<Column>) - 类 中的静态方法org.apache.spark.sql.expressions.Window
-
Creates a
WindowSpec with the partitioning defined.
- partitionBy(String, String...) - 类 中的方法org.apache.spark.sql.expressions.WindowSpec
-
- partitionBy(Column...) - 类 中的方法org.apache.spark.sql.expressions.WindowSpec
-
- partitionBy(String, Seq<String>) - 类 中的方法org.apache.spark.sql.expressions.WindowSpec
-
- partitionBy(Seq<Column>) - 类 中的方法org.apache.spark.sql.expressions.WindowSpec
-
- partitionBy(String...) - 类 中的方法org.apache.spark.sql.streaming.DataStreamWriter
-
Partitions the output by the given columns on the file system.
- partitionBy(Seq<String>) - 类 中的方法org.apache.spark.sql.streaming.DataStreamWriter
-
Partitions the output by the given columns on the file system.
- PartitionCoalescer - org.apache.spark.rdd中的接口
-
::DeveloperApi::
A PartitionCoalescer defines how to coalesce the partitions of a given RDD.
- partitioner() - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- partitioner() - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- partitioner() - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- partitioner() - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
The partitioner of this RDD.
- partitioner() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- partitioner() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- partitioner() - 类 中的方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
If partitionsRDD already has a partitioner, use it.
- partitioner() - 类 中的方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- partitioner() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- Partitioner - org.apache.spark中的类
-
An object that defines how the elements in a key-value pair RDD are partitioned by key.
- Partitioner() - 类 的构造器org.apache.spark.Partitioner
-
- partitioner() - 类 中的方法org.apache.spark.rdd.CoGroupedRDD
-
- partitioner() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- partitioner() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- partitioner() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- partitioner() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- partitioner() - 类 中的方法org.apache.spark.rdd.RDD
-
Optionally overridden by subclasses to specify how they are partitioned.
- partitioner() - 类 中的方法org.apache.spark.rdd.ShuffledRDD
-
- partitioner() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- partitioner() - 类 中的方法org.apache.spark.ShuffleDependency
-
- partitioner(Partitioner) - 类 中的方法org.apache.spark.streaming.StateSpec
-
Set the partitioner by which the state RDDs generated by mapWithState will be partitioned.
- PartitionGroup - org.apache.spark.rdd中的类
-
::DeveloperApi::
A group of Partitions
param: prefLoc preferred location for the partition group
- PartitionGroup(Option<String>) - 类 的构造器org.apache.spark.rdd.PartitionGroup
-
- partitionID() - 类 中的方法org.apache.spark.TaskCommitDenied
-
- partitionId() - 类 中的方法org.apache.spark.TaskContext
-
The ID of the RDD partition that is computed by this task.
- Partitioning - org.apache.spark.sql.sources.v2.reader.partitioning中的接口
-
- PartitionLocations(RDD<?>) - 类 的构造器org.apache.spark.rdd.DefaultPartitionCoalescer.PartitionLocations
-
- PartitionOffset - org.apache.spark.sql.sources.v2.reader.streaming中的接口
-
Used for per-partition offsets in continuous processing.
- PartitionPruningRDD<T> - org.apache.spark.rdd中的类
-
:: DeveloperApi ::
An RDD used to prune RDD partitions/partitions so we can avoid launching tasks on
all partitions.
- PartitionPruningRDD(RDD<T>, Function1<Object, Object>, ClassTag<T>) - 类 的构造器org.apache.spark.rdd.PartitionPruningRDD
-
- partitions() - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- partitions() - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- partitions() - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- partitions() - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Set of partitions in this RDD.
- partitions() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- partitions() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- partitions() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- partitions() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- partitions() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- partitions() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- partitions() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- partitions() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- partitions() - 类 中的方法org.apache.spark.rdd.PartitionGroup
-
- partitions() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- partitions() - 类 中的方法org.apache.spark.rdd.RDD
-
Get the array of partitions of this RDD, taking into account whether the
RDD is checkpointed or not.
- partitions() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- partitions() - 类 中的方法org.apache.spark.status.api.v1.RDDStorageInfo
-
- partitionsRDD() - 类 中的方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- partitionsRDD() - 类 中的方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- PartitionStrategy - org.apache.spark.graphx中的接口
-
Represents the way edges are assigned to edge partitions based on their source and destination
vertex IDs.
- PartitionStrategy.CanonicalRandomVertexCut$ - org.apache.spark.graphx中的类
-
Assigns edges to partitions by hashing the source and destination vertex IDs in a canonical
direction, resulting in a random vertex cut that colocates all edges between two vertices,
regardless of direction.
- PartitionStrategy.EdgePartition1D$ - org.apache.spark.graphx中的类
-
Assigns edges to partitions using only the source vertex ID, colocating edges with the same
source.
- PartitionStrategy.EdgePartition2D$ - org.apache.spark.graphx中的类
-
Assigns edges to partitions using a 2D partitioning of the sparse edge adjacency matrix,
guaranteeing a 2 * sqrt(numParts) bound on vertex replication.
- PartitionStrategy.RandomVertexCut$ - org.apache.spark.graphx中的类
-
Assigns edges to partitions by hashing the source and destination vertex IDs, resulting in a
random vertex cut that colocates all same-direction edges between two vertices.
- partsWithLocs() - 类 中的方法org.apache.spark.rdd.DefaultPartitionCoalescer.PartitionLocations
-
- partsWithoutLocs() - 类 中的方法org.apache.spark.rdd.DefaultPartitionCoalescer.PartitionLocations
-
- patch(int, GenSeq<B>, int, CanBuildFrom<Repr, B, That>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- path() - 类 中的方法org.apache.spark.scheduler.InputFormatInfo
-
- path() - 类 中的方法org.apache.spark.scheduler.SplitInfo
-
- pattern() - 类 中的方法org.apache.spark.ml.feature.RegexTokenizer
-
Regex pattern used to match delimiters if gaps is true or tokens if gaps is false.
- pc() - 类 中的方法org.apache.spark.ml.feature.PCAModel
-
- pc() - 类 中的方法org.apache.spark.mllib.feature.PCAModel
-
- PCA - org.apache.spark.ml.feature中的类
-
PCA trains a model to project vectors to a lower dimensional space of the top PCA!.
- PCA(String) - 类 的构造器org.apache.spark.ml.feature.PCA
-
- PCA() - 类 的构造器org.apache.spark.ml.feature.PCA
-
- PCA - org.apache.spark.mllib.feature中的类
-
A feature transformer that projects vectors to a low-dimensional space using PCA.
- PCA(int) - 类 的构造器org.apache.spark.mllib.feature.PCA
-
- PCAModel - org.apache.spark.ml.feature中的类
-
- PCAModel - org.apache.spark.mllib.feature中的类
-
Model fitted by
PCA that can project vectors to a low-dimensional space using PCA.
- PCAUtil - org.apache.spark.mllib.feature中的类
-
- PCAUtil() - 类 的构造器org.apache.spark.mllib.feature.PCAUtil
-
- pdf(Vector) - 类 中的方法org.apache.spark.ml.stat.distribution.MultivariateGaussian
-
Returns density of this multivariate Gaussian at given point, x
- pdf(Vector) - 类 中的方法org.apache.spark.mllib.stat.distribution.MultivariateGaussian
-
Returns density of this multivariate Gaussian at given point, x
- PEAK_EXECUTION_MEMORY() - 类 中的静态方法org.apache.spark.InternalAccumulator
-
- PEAK_EXECUTION_MEMORY() - 类 中的静态方法org.apache.spark.ui.jobs.TaskDetailsClassNames
-
- PEAK_EXECUTION_MEMORY() - 类 中的静态方法org.apache.spark.ui.ToolTips
-
- PEAK_MEM() - 类 中的静态方法org.apache.spark.status.TaskIndexNames
-
- peakExecutionMemory() - 类 中的方法org.apache.spark.status.api.v1.TaskMetricDistributions
-
- peakExecutionMemory() - 类 中的方法org.apache.spark.status.api.v1.TaskMetrics
-
- PEARSON() - 类 中的静态方法org.apache.spark.mllib.stat.test.ChiSqTest
-
- PearsonCorrelation - org.apache.spark.mllib.stat.correlation中的类
-
Compute Pearson correlation for two RDDs of the type RDD[Double] or the correlation matrix
for an RDD of the type RDD[Vector].
- PearsonCorrelation() - 类 的构造器org.apache.spark.mllib.stat.correlation.PearsonCorrelation
-
- percent_rank() - 类 中的静态方法org.apache.spark.sql.functions
-
Window function: returns the relative rank (i.e. percentile) of rows within a window partition.
- percentile() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelector
-
- percentile() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- percentile() - 类 中的方法org.apache.spark.mllib.feature.ChiSqSelector
-
- percentiles() - 类 中的静态方法org.apache.spark.scheduler.StatsReportListener
-
- percentilesHeader() - 类 中的静态方法org.apache.spark.scheduler.StatsReportListener
-
- permutations() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- persist(StorageLevel) - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
Set this RDD's storage level to persist its values across operations after the first time
it is computed.
- persist(StorageLevel) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Set this RDD's storage level to persist its values across operations after the first time
it is computed.
- persist(StorageLevel) - 类 中的方法org.apache.spark.api.java.JavaRDD
-
Set this RDD's storage level to persist its values across operations after the first time
it is computed.
- persist(StorageLevel) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- persist() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- persist(StorageLevel) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- persist() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- persist(StorageLevel) - 类 中的方法org.apache.spark.graphx.Graph
-
Caches the vertices and edges associated with this graph at the specified storage level,
ignoring any target storage levels previously set.
- persist(StorageLevel) - 类 中的方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
Persists the edge partitions at the specified storage level, ignoring any existing target
storage level.
- persist(StorageLevel) - 类 中的方法org.apache.spark.graphx.impl.GraphImpl
-
- persist(StorageLevel) - 类 中的方法org.apache.spark.graphx.impl.VertexRDDImpl
-
Persists the vertex partitions at the specified storage level, ignoring any existing target
storage level.
- persist(StorageLevel) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- persist() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- persist(StorageLevel) - 类 中的方法org.apache.spark.mllib.linalg.distributed.BlockMatrix
-
Persists the underlying RDD with the specified storage level.
- persist(StorageLevel) - 类 中的方法org.apache.spark.rdd.HadoopRDD
-
- persist(StorageLevel) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- persist() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- persist(StorageLevel) - 类 中的方法org.apache.spark.rdd.NewHadoopRDD
-
- persist(StorageLevel) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- persist() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- persist(StorageLevel) - 类 中的方法org.apache.spark.rdd.RDD
-
Set this RDD's storage level to persist its values across operations after the first time
it is computed.
- persist() - 类 中的方法org.apache.spark.rdd.RDD
-
Persist this RDD with the default storage level (MEMORY_ONLY).
- persist(StorageLevel) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- persist() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- persist() - 类 中的方法org.apache.spark.sql.Dataset
-
Persist this Dataset with the default storage level (MEMORY_AND_DISK).
- persist(StorageLevel) - 类 中的方法org.apache.spark.sql.Dataset
-
Persist this Dataset with the given storage level.
- persist() - 类 中的方法org.apache.spark.streaming.api.java.JavaDStream
-
Persist RDDs of this DStream with the default storage level (MEMORY_ONLY_SER)
- persist(StorageLevel) - 类 中的方法org.apache.spark.streaming.api.java.JavaDStream
-
Persist the RDDs of this DStream with the given storage level
- persist() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- persist(StorageLevel) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- persist() - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Persist RDDs of this DStream with the default storage level (MEMORY_ONLY_SER)
- persist(StorageLevel) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Persist the RDDs of this DStream with the given storage level
- persist() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- persist(StorageLevel) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- persist() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- persist(StorageLevel) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- persist() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- persist(StorageLevel) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- persist(StorageLevel) - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Persist the RDDs of this DStream with the given storage level
- persist() - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Persist RDDs of this DStream with the default storage level (MEMORY_ONLY_SER)
- persist$default$1() - 类 中的静态方法org.apache.spark.graphx.impl.GraphImpl
-
- personalizedPageRank(long, double, double) - 类 中的方法org.apache.spark.graphx.GraphOps
-
Run personalized PageRank for a given vertex, such that all random walks
are started relative to the source node.
- phrase(Parsers.Parser<T>) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaParser
-
- pi() - 类 中的方法org.apache.spark.ml.classification.NaiveBayesModel
-
- pi() - 类 中的方法org.apache.spark.mllib.classification.NaiveBayesModel
-
- pi() - 类 中的方法org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV1_0$.Data
-
- pi() - 类 中的方法org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV2_0$.Data
-
- pickBin(Partition, RDD<?>, double, DefaultPartitionCoalescer.PartitionLocations) - 类 中的方法org.apache.spark.rdd.DefaultPartitionCoalescer
-
Takes a parent RDD partition and decides which of the partition groups to put it in
Takes locality into account, but also uses power of 2 choices to load balance
It strikes a balance between the two using the balanceSlack variable
- pickRandomVertex() - 类 中的方法org.apache.spark.graphx.GraphOps
-
Picks a random vertex from the graph and returns its ID.
- pipe(String) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- pipe(List<String>) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- pipe(List<String>, Map<String, String>) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- pipe(List<String>, Map<String, String>, boolean, int) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- pipe(List<String>, Map<String, String>, boolean, int, String) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- pipe(String) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- pipe(List<String>) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- pipe(List<String>, Map<String, String>) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- pipe(List<String>, Map<String, String>, boolean, int) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- pipe(List<String>, Map<String, String>, boolean, int, String) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- pipe(String) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- pipe(List<String>) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- pipe(List<String>, Map<String, String>) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- pipe(List<String>, Map<String, String>, boolean, int) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- pipe(List<String>, Map<String, String>, boolean, int, String) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- pipe(String) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Return an RDD created by piping elements to a forked external process.
- pipe(List<String>) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Return an RDD created by piping elements to a forked external process.
- pipe(List<String>, Map<String, String>) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Return an RDD created by piping elements to a forked external process.
- pipe(List<String>, Map<String, String>, boolean, int) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Return an RDD created by piping elements to a forked external process.
- pipe(List<String>, Map<String, String>, boolean, int, String) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Return an RDD created by piping elements to a forked external process.
- pipe(String) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- pipe(String, Map<String, String>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- pipe(Seq<String>, Map<String, String>, Function1<Function1<String, BoxedUnit>, BoxedUnit>, Function2<T, Function1<String, BoxedUnit>, BoxedUnit>, boolean, int, String) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- pipe(String) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- pipe(String, Map<String, String>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- pipe(Seq<String>, Map<String, String>, Function1<Function1<String, BoxedUnit>, BoxedUnit>, Function2<T, Function1<String, BoxedUnit>, BoxedUnit>, boolean, int, String) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- pipe(String) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- pipe(String, Map<String, String>) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- pipe(Seq<String>, Map<String, String>, Function1<Function1<String, BoxedUnit>, BoxedUnit>, Function2<T, Function1<String, BoxedUnit>, BoxedUnit>, boolean, int, String) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- pipe(String) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- pipe(String, Map<String, String>) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- pipe(Seq<String>, Map<String, String>, Function1<Function1<String, BoxedUnit>, BoxedUnit>, Function2<T, Function1<String, BoxedUnit>, BoxedUnit>, boolean, int, String) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- pipe(String) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- pipe(String, Map<String, String>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- pipe(Seq<String>, Map<String, String>, Function1<Function1<String, BoxedUnit>, BoxedUnit>, Function2<T, Function1<String, BoxedUnit>, BoxedUnit>, boolean, int, String) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- pipe(String) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- pipe(String, Map<String, String>) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- pipe(Seq<String>, Map<String, String>, Function1<Function1<String, BoxedUnit>, BoxedUnit>, Function2<T, Function1<String, BoxedUnit>, BoxedUnit>, boolean, int, String) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- pipe(String) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- pipe(String, Map<String, String>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- pipe(Seq<String>, Map<String, String>, Function1<Function1<String, BoxedUnit>, BoxedUnit>, Function2<T, Function1<String, BoxedUnit>, BoxedUnit>, boolean, int, String) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- pipe(String) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- pipe(String, Map<String, String>) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- pipe(Seq<String>, Map<String, String>, Function1<Function1<String, BoxedUnit>, BoxedUnit>, Function2<T, Function1<String, BoxedUnit>, BoxedUnit>, boolean, int, String) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- pipe(String) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- pipe(String, Map<String, String>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- pipe(Seq<String>, Map<String, String>, Function1<Function1<String, BoxedUnit>, BoxedUnit>, Function2<T, Function1<String, BoxedUnit>, BoxedUnit>, boolean, int, String) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- pipe(String) - 类 中的方法org.apache.spark.rdd.RDD
-
Return an RDD created by piping elements to a forked external process.
- pipe(String, Map<String, String>) - 类 中的方法org.apache.spark.rdd.RDD
-
Return an RDD created by piping elements to a forked external process.
- pipe(Seq<String>, Map<String, String>, Function1<Function1<String, BoxedUnit>, BoxedUnit>, Function2<T, Function1<String, BoxedUnit>, BoxedUnit>, boolean, int, String) - 类 中的方法org.apache.spark.rdd.RDD
-
Return an RDD created by piping elements to a forked external process.
- pipe(String) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- pipe(String, Map<String, String>) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- pipe(Seq<String>, Map<String, String>, Function1<Function1<String, BoxedUnit>, BoxedUnit>, Function2<T, Function1<String, BoxedUnit>, BoxedUnit>, boolean, int, String) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- pipe$default$2() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- pipe$default$2() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- pipe$default$2() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- pipe$default$2() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- pipe$default$2() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- pipe$default$2() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- pipe$default$2() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- pipe$default$2() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- pipe$default$2() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- pipe$default$2() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- pipe$default$3() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- pipe$default$3() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- pipe$default$3() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- pipe$default$3() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- pipe$default$3() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- pipe$default$3() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- pipe$default$3() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- pipe$default$3() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- pipe$default$3() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- pipe$default$3() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- pipe$default$4() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- pipe$default$4() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- pipe$default$4() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- pipe$default$4() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- pipe$default$4() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- pipe$default$4() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- pipe$default$4() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- pipe$default$4() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- pipe$default$4() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- pipe$default$4() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- pipe$default$5() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- pipe$default$5() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- pipe$default$5() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- pipe$default$5() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- pipe$default$5() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- pipe$default$5() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- pipe$default$5() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- pipe$default$5() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- pipe$default$5() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- pipe$default$5() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- pipe$default$6() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- pipe$default$6() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- pipe$default$6() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- pipe$default$6() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- pipe$default$6() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- pipe$default$6() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- pipe$default$6() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- pipe$default$6() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- pipe$default$6() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- pipe$default$6() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- pipe$default$7() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- pipe$default$7() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- pipe$default$7() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- pipe$default$7() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- pipe$default$7() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- pipe$default$7() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- pipe$default$7() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- pipe$default$7() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- pipe$default$7() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- pipe$default$7() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- Pipeline - org.apache.spark.ml中的类
-
A simple pipeline, which acts as an estimator.
- Pipeline(String) - 类 的构造器org.apache.spark.ml.Pipeline
-
- Pipeline() - 类 的构造器org.apache.spark.ml.Pipeline
-
- Pipeline.SharedReadWrite$ - org.apache.spark.ml中的类
-
- PipelineModel - org.apache.spark.ml中的类
-
Represents a fitted pipeline.
- PipelineStage - org.apache.spark.ml中的类
-
- PipelineStage() - 类 的构造器org.apache.spark.ml.PipelineStage
-
- pivot(String) - 类 中的方法org.apache.spark.sql.RelationalGroupedDataset
-
Pivots a column of the current DataFrame and performs the specified aggregation.
- pivot(String, Seq<Object>) - 类 中的方法org.apache.spark.sql.RelationalGroupedDataset
-
Pivots a column of the current DataFrame and performs the specified aggregation.
- pivot(String, List<Object>) - 类 中的方法org.apache.spark.sql.RelationalGroupedDataset
-
(Java-specific) Pivots a column of the current DataFrame and performs the specified
aggregation.
- PivotType$() - 类 的构造器org.apache.spark.sql.RelationalGroupedDataset.PivotType$
-
- plan() - 异常错误 中的方法org.apache.spark.sql.AnalysisException
-
- plus(Object) - 类 中的方法org.apache.spark.sql.Column
-
Sum of this expression and another expression.
- plus(Duration) - 类 中的方法org.apache.spark.streaming.Duration
-
- plus(Duration) - 类 中的方法org.apache.spark.streaming.Time
-
- PMMLExportable - org.apache.spark.mllib.pmml中的接口
-
:: DeveloperApi ::
Export model to the PMML format
Predictive Model Markup Language (PMML) is an XML-based file format
developed by the Data Mining Group (www.dmg.org).
- PMMLModelExportFactory - org.apache.spark.mllib.pmml.export中的类
-
- PMMLModelExportFactory() - 类 的构造器org.apache.spark.mllib.pmml.export.PMMLModelExportFactory
-
- pmod(Column, Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the positive value of dividend mod divisor.
- point() - 类 中的方法org.apache.spark.mllib.feature.VocabWord
-
- POINTS() - 类 中的静态方法org.apache.spark.mllib.clustering.StreamingKMeans
-
- POISON_PILL() - 类 中的静态方法org.apache.spark.scheduler.AsyncEventQueue
-
- Poisson$() - 类 的构造器org.apache.spark.ml.regression.GeneralizedLinearRegression.Poisson$
-
- PoissonBounds - org.apache.spark.util.random中的类
-
Utility functions that help us determine bounds on adjusted sampling rate to guarantee exact
sample sizes with high confidence when sampling with replacement.
- PoissonBounds() - 类 的构造器org.apache.spark.util.random.PoissonBounds
-
- PoissonGenerator - org.apache.spark.mllib.random中的类
-
:: DeveloperApi ::
Generates i.i.d. samples from the Poisson distribution with the given mean.
- PoissonGenerator(double) - 类 的构造器org.apache.spark.mllib.random.PoissonGenerator
-
- poissonJavaRDD(JavaSparkContext, double, long, int, long) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
Java-friendly version of RandomRDDs.poissonRDD.
- poissonJavaRDD(JavaSparkContext, double, long, int) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
RandomRDDs.poissonJavaRDD with the default seed.
- poissonJavaRDD(JavaSparkContext, double, long) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
RandomRDDs.poissonJavaRDD with the default number of partitions and the default seed.
- poissonJavaVectorRDD(JavaSparkContext, double, long, int, int, long) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
Java-friendly version of RandomRDDs.poissonVectorRDD.
- poissonJavaVectorRDD(JavaSparkContext, double, long, int, int) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
RandomRDDs.poissonJavaVectorRDD with the default seed.
- poissonJavaVectorRDD(JavaSparkContext, double, long, int) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
RandomRDDs.poissonJavaVectorRDD with the default number of partitions and the default seed.
- poissonRDD(SparkContext, double, long, int, long) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
Generates an RDD comprised of i.i.d.
- PoissonSampler<T> - org.apache.spark.util.random中的类
-
:: DeveloperApi ::
A sampler for sampling with replacement, based on values drawn from Poisson distribution.
- PoissonSampler(double, boolean) - 类 的构造器org.apache.spark.util.random.PoissonSampler
-
- PoissonSampler(double) - 类 的构造器org.apache.spark.util.random.PoissonSampler
-
- poissonVectorRDD(SparkContext, double, long, int, int, long) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
Generates an RDD[Vector] with vectors containing i.i.d.
- PolynomialExpansion - org.apache.spark.ml.feature中的类
-
Perform feature expansion in a polynomial space.
- PolynomialExpansion(String) - 类 的构造器org.apache.spark.ml.feature.PolynomialExpansion
-
- PolynomialExpansion() - 类 的构造器org.apache.spark.ml.feature.PolynomialExpansion
-
- popStdev() - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
Compute the population standard deviation of this RDD's elements.
- popStdev() - 类 中的方法org.apache.spark.rdd.DoubleRDDFunctions
-
Compute the population standard deviation of this RDD's elements.
- popStdev() - 类 中的方法org.apache.spark.util.StatCounter
-
Return the population standard deviation of the values.
- popVariance() - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
Compute the population variance of this RDD's elements.
- popVariance() - 类 中的方法org.apache.spark.rdd.DoubleRDDFunctions
-
Compute the population variance of this RDD's elements.
- popVariance() - 类 中的方法org.apache.spark.util.StatCounter
-
Return the population variance of the values.
- port() - 接口 中的方法org.apache.spark.SparkExecutorInfo
-
- port() - 类 中的方法org.apache.spark.SparkExecutorInfoImpl
-
- port() - 类 中的方法org.apache.spark.storage.BlockManagerId
-
- PortableDataStream - org.apache.spark.input中的类
-
A class that allows DataStreams to be serialized and moved around by not creating them
until they need to be read
- PortableDataStream(CombineFileSplit, TaskAttemptContext, Integer) - 类 的构造器org.apache.spark.input.PortableDataStream
-
- portMaxRetries(SparkConf) - 类 中的静态方法org.apache.spark.util.Utils
-
Maximum number of retries when binding to a port before giving up.
- posexplode(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Creates a new row for each element with position in the given array or map column.
- posexplode_outer(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Creates a new row for each element with position in the given array or map column.
- position() - 类 中的方法org.apache.spark.storage.ReadableChannelFileRegion
-
- positioned(Function0<Parsers.Parser<T>>) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaParser
-
- post(SparkListenerEvent) - 类 中的方法org.apache.spark.scheduler.AsyncEventQueue
-
- Postfix$() - 类 的构造器org.apache.spark.mllib.fpm.PrefixSpan.Postfix$
-
- PostgresDialect - org.apache.spark.sql.jdbc中的类
-
- PostgresDialect() - 类 的构造器org.apache.spark.sql.jdbc.PostgresDialect
-
- postToAll(E) - 类 中的静态方法org.apache.spark.scheduler.AsyncEventQueue
-
- pow(Column, Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the value of the first argument raised to the power of the second argument.
- pow(Column, String) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the value of the first argument raised to the power of the second argument.
- pow(String, Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the value of the first argument raised to the power of the second argument.
- pow(String, String) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the value of the first argument raised to the power of the second argument.
- pow(Column, double) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the value of the first argument raised to the power of the second argument.
- pow(String, double) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the value of the first argument raised to the power of the second argument.
- pow(double, Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the value of the first argument raised to the power of the second argument.
- pow(double, String) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the value of the first argument raised to the power of the second argument.
- PowerIterationClustering - org.apache.spark.mllib.clustering中的类
-
Power Iteration Clustering (PIC), a scalable graph clustering algorithm developed by
Lin and Cohen.
- PowerIterationClustering() - 类 的构造器org.apache.spark.mllib.clustering.PowerIterationClustering
-
Constructs a PIC instance with default parameters: {k: 2, maxIterations: 100,
initMode: "random"}.
- PowerIterationClustering.Assignment - org.apache.spark.mllib.clustering中的类
-
Cluster assignment.
- PowerIterationClustering.Assignment$ - org.apache.spark.mllib.clustering中的类
-
- PowerIterationClusteringModel - org.apache.spark.mllib.clustering中的类
-
- PowerIterationClusteringModel(int, RDD<PowerIterationClustering.Assignment>) - 类 的构造器org.apache.spark.mllib.clustering.PowerIterationClusteringModel
-
- PowerIterationClusteringModel.SaveLoadV1_0$ - org.apache.spark.mllib.clustering中的类
-
- pr() - 接口 中的方法org.apache.spark.ml.classification.BinaryLogisticRegressionSummary
-
Returns the precision-recall curve, which is a Dataframe containing
two fields recall, precision with (0.0, 1.0) prepended to it.
- pr() - 类 中的方法org.apache.spark.mllib.evaluation.BinaryClassificationMetrics
-
Returns the precision-recall curve, which is an RDD of (recall, precision),
NOT (precision, recall), with (0.0, p) prepended to it, where p is the precision
associated with the lowest recall on the curve.
- Precision - org.apache.spark.mllib.evaluation.binary中的类
-
Precision.
- Precision() - 类 的构造器org.apache.spark.mllib.evaluation.binary.Precision
-
- precision(double) - 类 中的方法org.apache.spark.mllib.evaluation.MulticlassMetrics
-
Returns precision for a given label (category)
- precision() - 类 中的方法org.apache.spark.mllib.evaluation.MulticlassMetrics
-
- precision() - 类 中的方法org.apache.spark.mllib.evaluation.MultilabelMetrics
-
Returns document-based precision averaged by the number of documents
- precision(double) - 类 中的方法org.apache.spark.mllib.evaluation.MultilabelMetrics
-
Returns precision for a given label (category)
- precision() - 类 中的方法org.apache.spark.sql.types.Decimal
-
- precision() - 类 中的方法org.apache.spark.sql.types.DecimalType
-
- precisionAt(int) - 类 中的方法org.apache.spark.mllib.evaluation.RankingMetrics
-
Compute the average precision of all the queries, truncated at ranking position k.
- precisionByLabel() - 接口 中的方法org.apache.spark.ml.classification.LogisticRegressionSummary
-
Returns precision for each label (category).
- precisionByThreshold() - 接口 中的方法org.apache.spark.ml.classification.BinaryLogisticRegressionSummary
-
Returns a dataframe with two fields (threshold, precision) curve.
- precisionByThreshold() - 类 中的方法org.apache.spark.mllib.evaluation.BinaryClassificationMetrics
-
Returns the (threshold, precision) curve.
- predict(Vector) - 类 中的方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- predict(RDD<Vector>) - 接口 中的方法org.apache.spark.mllib.classification.ClassificationModel
-
Predict values for the given data set using the model trained.
- predict(Vector) - 接口 中的方法org.apache.spark.mllib.classification.ClassificationModel
-
Predict values for a single data point using the model trained.
- predict(JavaRDD<Vector>) - 接口 中的方法org.apache.spark.mllib.classification.ClassificationModel
-
Predict values for examples stored in a JavaRDD.
- predict(RDD<Vector>) - 类 中的静态方法org.apache.spark.mllib.classification.LogisticRegressionModel
-
- predict(Vector) - 类 中的静态方法org.apache.spark.mllib.classification.LogisticRegressionModel
-
- predict(JavaRDD<Vector>) - 类 中的静态方法org.apache.spark.mllib.classification.LogisticRegressionModel
-
- predict(RDD<Vector>) - 类 中的方法org.apache.spark.mllib.classification.NaiveBayesModel
-
- predict(Vector) - 类 中的方法org.apache.spark.mllib.classification.NaiveBayesModel
-
- predict(RDD<Vector>) - 类 中的静态方法org.apache.spark.mllib.classification.SVMModel
-
- predict(Vector) - 类 中的静态方法org.apache.spark.mllib.classification.SVMModel
-
- predict(JavaRDD<Vector>) - 类 中的静态方法org.apache.spark.mllib.classification.SVMModel
-
- predict(Vector) - 类 中的方法org.apache.spark.mllib.clustering.BisectingKMeansModel
-
Predicts the index of the cluster that the input point belongs to.
- predict(RDD<Vector>) - 类 中的方法org.apache.spark.mllib.clustering.BisectingKMeansModel
-
Predicts the indices of the clusters that the input points belong to.
- predict(JavaRDD<Vector>) - 类 中的方法org.apache.spark.mllib.clustering.BisectingKMeansModel
-
Java-friendly version of predict().
- predict(RDD<Vector>) - 类 中的方法org.apache.spark.mllib.clustering.GaussianMixtureModel
-
Maps given points to their cluster indices.
- predict(Vector) - 类 中的方法org.apache.spark.mllib.clustering.GaussianMixtureModel
-
Maps given point to its cluster index.
- predict(JavaRDD<Vector>) - 类 中的方法org.apache.spark.mllib.clustering.GaussianMixtureModel
-
Java-friendly version of predict()
- predict(Vector) - 类 中的方法org.apache.spark.mllib.clustering.KMeansModel
-
Returns the cluster index that a given point belongs to.
- predict(RDD<Vector>) - 类 中的方法org.apache.spark.mllib.clustering.KMeansModel
-
Maps given points to their cluster indices.
- predict(JavaRDD<Vector>) - 类 中的方法org.apache.spark.mllib.clustering.KMeansModel
-
Maps given points to their cluster indices.
- predict(int, int) - 类 中的方法org.apache.spark.mllib.recommendation.MatrixFactorizationModel
-
Predict the rating of one user for one product.
- predict(RDD<Tuple2<Object, Object>>) - 类 中的方法org.apache.spark.mllib.recommendation.MatrixFactorizationModel
-
Predict the rating of many users for many products.
- predict(JavaPairRDD<Integer, Integer>) - 类 中的方法org.apache.spark.mllib.recommendation.MatrixFactorizationModel
-
Java-friendly version of MatrixFactorizationModel.predict.
- predict(RDD<Vector>) - 类 中的方法org.apache.spark.mllib.regression.GeneralizedLinearModel
-
Predict values for the given data set using the model trained.
- predict(Vector) - 类 中的方法org.apache.spark.mllib.regression.GeneralizedLinearModel
-
Predict values for a single data point using the model trained.
- predict(RDD<Object>) - 类 中的方法org.apache.spark.mllib.regression.IsotonicRegressionModel
-
Predict labels for provided features.
- predict(JavaDoubleRDD) - 类 中的方法org.apache.spark.mllib.regression.IsotonicRegressionModel
-
Predict labels for provided features.
- predict(double) - 类 中的方法org.apache.spark.mllib.regression.IsotonicRegressionModel
-
Predict a single label.
- predict(RDD<Vector>) - 类 中的静态方法org.apache.spark.mllib.regression.LassoModel
-
- predict(Vector) - 类 中的静态方法org.apache.spark.mllib.regression.LassoModel
-
- predict(JavaRDD<Vector>) - 类 中的静态方法org.apache.spark.mllib.regression.LassoModel
-
- predict(RDD<Vector>) - 类 中的静态方法org.apache.spark.mllib.regression.LinearRegressionModel
-
- predict(Vector) - 类 中的静态方法org.apache.spark.mllib.regression.LinearRegressionModel
-
- predict(JavaRDD<Vector>) - 类 中的静态方法org.apache.spark.mllib.regression.LinearRegressionModel
-
- predict(RDD<Vector>) - 接口 中的方法org.apache.spark.mllib.regression.RegressionModel
-
Predict values for the given data set using the model trained.
- predict(Vector) - 接口 中的方法org.apache.spark.mllib.regression.RegressionModel
-
Predict values for a single data point using the model trained.
- predict(JavaRDD<Vector>) - 接口 中的方法org.apache.spark.mllib.regression.RegressionModel
-
Predict values for examples stored in a JavaRDD.
- predict(RDD<Vector>) - 类 中的静态方法org.apache.spark.mllib.regression.RidgeRegressionModel
-
- predict(Vector) - 类 中的静态方法org.apache.spark.mllib.regression.RidgeRegressionModel
-
- predict(JavaRDD<Vector>) - 类 中的静态方法org.apache.spark.mllib.regression.RidgeRegressionModel
-
- predict(Vector) - 类 中的方法org.apache.spark.mllib.tree.model.DecisionTreeModel
-
Predict values for a single data point using the model trained.
- predict(RDD<Vector>) - 类 中的方法org.apache.spark.mllib.tree.model.DecisionTreeModel
-
Predict values for the given data set using the model trained.
- predict(JavaRDD<Vector>) - 类 中的方法org.apache.spark.mllib.tree.model.DecisionTreeModel
-
Predict values for the given data set using the model trained.
- predict() - 类 中的方法org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.NodeData
-
- predict() - 类 中的方法org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.PredictData
-
- predict(Vector) - 类 中的静态方法org.apache.spark.mllib.tree.model.GradientBoostedTreesModel
-
- predict(RDD<Vector>) - 类 中的静态方法org.apache.spark.mllib.tree.model.GradientBoostedTreesModel
-
- predict(JavaRDD<Vector>) - 类 中的静态方法org.apache.spark.mllib.tree.model.GradientBoostedTreesModel
-
- predict() - 类 中的方法org.apache.spark.mllib.tree.model.Node
-
- predict(Vector) - 类 中的方法org.apache.spark.mllib.tree.model.Node
-
predict value if node is not leaf
- Predict - org.apache.spark.mllib.tree.model中的类
-
:: DeveloperApi ::
Predicted value for a node
param: predict predicted value
param: prob probability of the label (classification only)
- Predict(double, double) - 类 的构造器org.apache.spark.mllib.tree.model.Predict
-
- predict() - 类 中的方法org.apache.spark.mllib.tree.model.Predict
-
- predict(Vector) - 类 中的静态方法org.apache.spark.mllib.tree.model.RandomForestModel
-
- predict(RDD<Vector>) - 类 中的静态方法org.apache.spark.mllib.tree.model.RandomForestModel
-
- predict(JavaRDD<Vector>) - 类 中的静态方法org.apache.spark.mllib.tree.model.RandomForestModel
-
- PredictData(double, double) - 类 的构造器org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.PredictData
-
- prediction() - 类 中的方法org.apache.spark.ml.tree.DecisionTreeModelReadWrite.NodeData
-
- prediction() - 类 中的方法org.apache.spark.ml.tree.InternalNode
-
- prediction() - 类 中的方法org.apache.spark.ml.tree.LeafNode
-
- prediction() - 类 中的方法org.apache.spark.ml.tree.Node
-
Prediction a leaf node makes, or which an internal node would make if it were a leaf node
- predictionCol() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- predictionCol() - 接口 中的方法org.apache.spark.ml.classification.LogisticRegressionSummary
-
Field in "predictions" which gives the prediction of each class.
- predictionCol() - 类 中的方法org.apache.spark.ml.classification.LogisticRegressionSummaryImpl
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.classification.OneVsRest
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.classification.OneVsRestModel
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeans
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
- predictionCol() - 类 中的方法org.apache.spark.ml.clustering.ClusteringSummary
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixture
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.clustering.KMeans
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.clustering.KMeansModel
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.evaluation.ClusteringEvaluator
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.evaluation.RegressionEvaluator
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowth
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowthModel
-
- predictionCol() - 接口 中的方法org.apache.spark.ml.param.shared.HasPredictionCol
-
Param for prediction column name.
- predictionCol() - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.recommendation.ALSModel
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- predictionCol() - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegressionSummary
-
Field in "predictions" which gives the predicted value of each instance.
- predictionCol() - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegression
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegressionModel
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- predictionCol() - 类 中的方法org.apache.spark.ml.regression.LinearRegressionSummary
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- predictionCol() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- PredictionModel<FeaturesType,M extends PredictionModel<FeaturesType,M>> - org.apache.spark.ml中的类
-
:: DeveloperApi ::
Abstraction for a model for prediction tasks (regression and classification).
- PredictionModel() - 类 的构造器org.apache.spark.ml.PredictionModel
-
- predictions() - 接口 中的方法org.apache.spark.ml.classification.LogisticRegressionSummary
-
Dataframe output by the model's transform method.
- predictions() - 类 中的方法org.apache.spark.ml.classification.LogisticRegressionSummaryImpl
-
- predictions() - 类 中的方法org.apache.spark.ml.clustering.ClusteringSummary
-
- predictions() - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegressionSummary
-
Predictions output by the model's transform method.
- predictions() - 类 中的方法org.apache.spark.ml.regression.IsotonicRegressionModel
-
Predictions associated with the boundaries at the same index, monotone because of isotonic
regression.
- predictions() - 类 中的方法org.apache.spark.ml.regression.LinearRegressionSummary
-
- predictions() - 类 中的方法org.apache.spark.mllib.regression.IsotonicRegressionModel
-
- predictOn(DStream<Vector>) - 类 中的方法org.apache.spark.mllib.clustering.StreamingKMeans
-
Use the clustering model to make predictions on batches of data from a DStream.
- predictOn(JavaDStream<Vector>) - 类 中的方法org.apache.spark.mllib.clustering.StreamingKMeans
-
Java-friendly version of predictOn.
- predictOn(DStream<Vector>) - 类 中的方法org.apache.spark.mllib.regression.StreamingLinearAlgorithm
-
Use the model to make predictions on batches of data from a DStream
- predictOn(JavaDStream<Vector>) - 类 中的方法org.apache.spark.mllib.regression.StreamingLinearAlgorithm
-
Java-friendly version of predictOn.
- predictOnValues(DStream<Tuple2<K, Vector>>, ClassTag<K>) - 类 中的方法org.apache.spark.mllib.clustering.StreamingKMeans
-
Use the model to make predictions on the values of a DStream and carry over its keys.
- predictOnValues(JavaPairDStream<K, Vector>) - 类 中的方法org.apache.spark.mllib.clustering.StreamingKMeans
-
Java-friendly version of predictOnValues.
- predictOnValues(DStream<Tuple2<K, Vector>>, ClassTag<K>) - 类 中的方法org.apache.spark.mllib.regression.StreamingLinearAlgorithm
-
Use the model to make predictions on the values of a DStream and carry over its keys.
- predictOnValues(JavaPairDStream<K, Vector>) - 类 中的方法org.apache.spark.mllib.regression.StreamingLinearAlgorithm
-
Java-friendly version of predictOnValues.
- Predictor<FeaturesType,Learner extends Predictor<FeaturesType,Learner,M>,M extends PredictionModel<FeaturesType,M>> - org.apache.spark.ml中的类
-
:: DeveloperApi ::
Abstraction for prediction problems (regression and classification).
- Predictor() - 类 的构造器org.apache.spark.ml.Predictor
-
- predictProbabilities(RDD<Vector>) - 类 中的方法org.apache.spark.mllib.classification.NaiveBayesModel
-
Predict values for the given data set using the model trained.
- predictProbabilities(Vector) - 类 中的方法org.apache.spark.mllib.classification.NaiveBayesModel
-
Predict posterior class probabilities for a single data point using the model trained.
- predictQuantiles(Vector) - 类 中的方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- predictSoft(RDD<Vector>) - 类 中的方法org.apache.spark.mllib.clustering.GaussianMixtureModel
-
Given the input vectors, return the membership value of each vector
to all mixture components.
- predictSoft(Vector) - 类 中的方法org.apache.spark.mllib.clustering.GaussianMixtureModel
-
Given the input vector, return the membership values to all mixture components.
- preferredLocation() - 类 中的方法org.apache.spark.streaming.receiver.Receiver
-
Override this to specify a preferred location (hostname).
- preferredLocations(Partition) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- preferredLocations(Partition) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- preferredLocations(Partition) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- preferredLocations(Partition) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- preferredLocations(Partition) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- preferredLocations(Partition) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- preferredLocations(Partition) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- preferredLocations(Partition) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- preferredLocations(Partition) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- preferredLocations(Partition) - 类 中的方法org.apache.spark.rdd.RDD
-
Get the preferred locations of a partition, taking into account whether the
RDD is checkpointed.
- preferredLocations(Partition) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- preferredLocations() - 接口 中的方法org.apache.spark.sql.sources.v2.reader.DataReaderFactory
-
The preferred locations where the data reader returned by this reader factory can run faster,
but Spark does not guarantee to run the data reader on these locations.
- Prefix$() - 类 的构造器org.apache.spark.mllib.fpm.PrefixSpan.Prefix$
-
- prefixesToRewrite() - 类 中的方法org.apache.spark.ml.feature.VectorAttributeRewriter
-
- prefixLength(Function1<A, Object>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- PrefixSpan - org.apache.spark.mllib.fpm中的类
-
A parallel PrefixSpan algorithm to mine frequent sequential patterns.
- PrefixSpan() - 类 的构造器org.apache.spark.mllib.fpm.PrefixSpan
-
Constructs a default instance with default parameters
{minSupport: 0.1, maxPatternLength: 10, maxLocalProjDBSize: 32000000L}.
- PrefixSpan.FreqSequence<Item> - org.apache.spark.mllib.fpm中的类
-
Represents a frequent sequence.
- PrefixSpan.Postfix$ - org.apache.spark.mllib.fpm中的类
-
- PrefixSpan.Prefix$ - org.apache.spark.mllib.fpm中的类
-
- PrefixSpanModel<Item> - org.apache.spark.mllib.fpm中的类
-
Model fitted by
PrefixSpan
param: freqSequences frequent sequences
- PrefixSpanModel(RDD<PrefixSpan.FreqSequence<Item>>) - 类 的构造器org.apache.spark.mllib.fpm.PrefixSpanModel
-
- PrefixSpanModel.SaveLoadV1_0$ - org.apache.spark.mllib.fpm中的类
-
- prefLoc() - 类 中的方法org.apache.spark.rdd.PartitionGroup
-
- pregel(A, int, EdgeDirection, Function3<Object, VD, A, VD>, Function1<EdgeTriplet<VD, ED>, Iterator<Tuple2<Object, A>>>, Function2<A, A, A>, ClassTag<A>) - 类 中的方法org.apache.spark.graphx.GraphOps
-
Execute a Pregel-like iterative vertex-parallel abstraction.
- Pregel - org.apache.spark.graphx中的类
-
Implements a Pregel-like bulk-synchronous message-passing API.
- Pregel() - 类 的构造器org.apache.spark.graphx.Pregel
-
- prepare() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- prepareWritable(Writable, Seq<Tuple2<String, String>>) - 类 中的静态方法org.apache.spark.sql.hive.HiveShim
-
- prepareWrite(SparkSession, Job, Map<String, String>, StructType) - 类 中的方法org.apache.spark.sql.hive.execution.HiveFileFormat
-
- prepareWrite(SparkSession, Job, Map<String, String>, StructType) - 类 中的方法org.apache.spark.sql.hive.orc.OrcFileFormat
-
- prependBaseUri(String, String) - 类 中的静态方法org.apache.spark.ui.UIUtils
-
- prettyJson() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- prettyJson() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- prettyJson() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- prettyJson() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- prettyJson() - 类 中的方法org.apache.spark.sql.streaming.SinkProgress
-
The pretty (i.e. indented) JSON representation of this progress.
- prettyJson() - 类 中的方法org.apache.spark.sql.streaming.SourceProgress
-
The pretty (i.e. indented) JSON representation of this progress.
- prettyJson() - 类 中的方法org.apache.spark.sql.streaming.StateOperatorProgress
-
The pretty (i.e. indented) JSON representation of this progress.
- prettyJson() - 类 中的方法org.apache.spark.sql.streaming.StreamingQueryProgress
-
The pretty (i.e. indented) JSON representation of this progress.
- prettyJson() - 类 中的方法org.apache.spark.sql.streaming.StreamingQueryStatus
-
The pretty (i.e. indented) JSON representation of this status.
- prettyJson() - 类 中的静态方法org.apache.spark.sql.types.ArrayType
-
- prettyJson() - 类 中的静态方法org.apache.spark.sql.types.BinaryType
-
- prettyJson() - 类 中的静态方法org.apache.spark.sql.types.BooleanType
-
- prettyJson() - 类 中的静态方法org.apache.spark.sql.types.ByteType
-
- prettyJson() - 类 中的静态方法org.apache.spark.sql.types.CalendarIntervalType
-
- prettyJson() - 类 中的静态方法org.apache.spark.sql.types.CharType
-
- prettyJson() - 类 中的方法org.apache.spark.sql.types.DataType
-
The pretty (i.e. indented) JSON representation of this data type.
- prettyJson() - 类 中的静态方法org.apache.spark.sql.types.DateType
-
- prettyJson() - 类 中的静态方法org.apache.spark.sql.types.DecimalType
-
- prettyJson() - 类 中的静态方法org.apache.spark.sql.types.DoubleType
-
- prettyJson() - 类 中的静态方法org.apache.spark.sql.types.FloatType
-
- prettyJson() - 类 中的静态方法org.apache.spark.sql.types.HiveStringType
-
- prettyJson() - 类 中的静态方法org.apache.spark.sql.types.IntegerType
-
- prettyJson() - 类 中的静态方法org.apache.spark.sql.types.LongType
-
- prettyJson() - 类 中的静态方法org.apache.spark.sql.types.MapType
-
- prettyJson() - 类 中的静态方法org.apache.spark.sql.types.NullType
-
- prettyJson() - 类 中的静态方法org.apache.spark.sql.types.NumericType
-
- prettyJson() - 类 中的静态方法org.apache.spark.sql.types.ObjectType
-
- prettyJson() - 类 中的静态方法org.apache.spark.sql.types.ShortType
-
- prettyJson() - 类 中的静态方法org.apache.spark.sql.types.StringType
-
- prettyJson() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- prettyJson() - 类 中的静态方法org.apache.spark.sql.types.TimestampType
-
- prettyJson() - 类 中的静态方法org.apache.spark.sql.types.VarcharType
-
- prettyPrint() - 类 中的方法org.apache.spark.streaming.Duration
-
- prev() - 类 中的方法org.apache.spark.rdd.ShuffledRDD
-
- prev() - 类 中的方法org.apache.spark.status.LiveRDDPartition
-
- print() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- print(int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- print() - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
Print the first ten elements of each RDD generated in this DStream.
- print(int) - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
Print the first num elements of each RDD generated in this DStream.
- print() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- print(int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- print() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- print(int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- print() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- print(int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- print() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- print(int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- print() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- print(int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- print() - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Print the first ten elements of each RDD generated in this DStream.
- print(int) - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Print the first num elements of each RDD generated in this DStream.
- printSchema() - 类 中的方法org.apache.spark.sql.Dataset
-
Prints the schema to the console in a nice tree format.
- printSchema() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- printSchema() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- printSchema() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- printSchema() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- printStackTrace() - 异常错误 中的静态方法org.apache.spark.sql.AnalysisException
-
- printStackTrace(PrintStream) - 异常错误 中的静态方法org.apache.spark.sql.AnalysisException
-
- printStackTrace(PrintWriter) - 异常错误 中的静态方法org.apache.spark.sql.AnalysisException
-
- printStats() - 类 中的方法org.apache.spark.streaming.scheduler.StatsReportListener
-
- printTreeString() - 类 中的方法org.apache.spark.sql.types.StructType
-
- prioritize(BlockManagerId, Seq<BlockManagerId>, HashSet<BlockManagerId>, BlockId, int) - 类 中的方法org.apache.spark.storage.BasicBlockReplicationPolicy
-
Method to prioritize a bunch of candidate peers of a block manager.
- prioritize(BlockManagerId, Seq<BlockManagerId>, HashSet<BlockManagerId>, BlockId, int) - 接口 中的方法org.apache.spark.storage.BlockReplicationPolicy
-
Method to prioritize a bunch of candidate peers of a block
- prioritize(BlockManagerId, Seq<BlockManagerId>, HashSet<BlockManagerId>, BlockId, int) - 类 中的方法org.apache.spark.storage.RandomBlockReplicationPolicy
-
Method to prioritize a bunch of candidate peers of a block.
- prob() - 类 中的方法org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.PredictData
-
- prob() - 类 中的方法org.apache.spark.mllib.tree.model.Predict
-
- ProbabilisticClassificationModel<FeaturesType,M extends ProbabilisticClassificationModel<FeaturesType,M>> - org.apache.spark.ml.classification中的类
-
- ProbabilisticClassificationModel() - 类 的构造器org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- ProbabilisticClassifier<FeaturesType,E extends ProbabilisticClassifier<FeaturesType,E,M>,M extends ProbabilisticClassificationModel<FeaturesType,M>> - org.apache.spark.ml.classification中的类
-
:: DeveloperApi ::
Single-label binary or multiclass classifier which can output class conditional probabilities.
- ProbabilisticClassifier() - 类 的构造器org.apache.spark.ml.classification.ProbabilisticClassifier
-
- probabilities() - 类 中的静态方法org.apache.spark.scheduler.StatsReportListener
-
- probability() - 类 中的方法org.apache.spark.ml.clustering.GaussianMixtureSummary
-
Probability of each cluster.
- probabilityCol() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- probabilityCol() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- probabilityCol() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- probabilityCol() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- probabilityCol() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- probabilityCol() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- probabilityCol() - 接口 中的方法org.apache.spark.ml.classification.LogisticRegressionSummary
-
Field in "predictions" which gives the probability of each class as a vector.
- probabilityCol() - 类 中的方法org.apache.spark.ml.classification.LogisticRegressionSummaryImpl
-
- probabilityCol() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- probabilityCol() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- probabilityCol() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- probabilityCol() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- probabilityCol() - 类 中的静态方法org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- probabilityCol() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- probabilityCol() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- probabilityCol() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixture
-
- probabilityCol() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- probabilityCol() - 类 中的方法org.apache.spark.ml.clustering.GaussianMixtureSummary
-
- probabilityCol() - 接口 中的方法org.apache.spark.ml.param.shared.HasProbabilityCol
-
Param for Column name for predicted class conditional probabilities.
- Probit$() - 类 的构造器org.apache.spark.ml.regression.GeneralizedLinearRegression.Probit$
-
- process(T) - 类 中的方法org.apache.spark.sql.ForeachWriter
-
Called to process the data in the executor side.
- PROCESS_LOCAL() - 类 中的静态方法org.apache.spark.scheduler.TaskLocality
-
- processAllAvailable() - 接口 中的方法org.apache.spark.sql.streaming.StreamingQuery
-
Blocks until all available data in the source has been processed and committed to the sink.
- processedRowsPerSecond() - 类 中的方法org.apache.spark.sql.streaming.SourceProgress
-
- processedRowsPerSecond() - 类 中的方法org.apache.spark.sql.streaming.StreamingQueryProgress
-
The aggregate (across all sources) rate at which Spark is processing data.
- processingDelay() - 类 中的方法org.apache.spark.streaming.scheduler.BatchInfo
-
Time taken for the all jobs of this batch to finish processing from the time they started
processing.
- processingEndTime() - 类 中的方法org.apache.spark.streaming.scheduler.BatchInfo
-
- processingStartTime() - 类 中的方法org.apache.spark.streaming.scheduler.BatchInfo
-
- ProcessingTime - org.apache.spark.sql.streaming中的类
-
- ProcessingTime(long) - 类 的构造器org.apache.spark.sql.streaming.ProcessingTime
-
已过时。
- ProcessingTime(long) - 类 中的静态方法org.apache.spark.sql.streaming.Trigger
-
A trigger policy that runs a query periodically based on an interval in processing time.
- ProcessingTime(long, TimeUnit) - 类 中的静态方法org.apache.spark.sql.streaming.Trigger
-
(Java-friendly)
A trigger policy that runs a query periodically based on an interval in processing time.
- ProcessingTime(Duration) - 类 中的静态方法org.apache.spark.sql.streaming.Trigger
-
(Scala-friendly)
A trigger policy that runs a query periodically based on an interval in processing time.
- ProcessingTime(String) - 类 中的静态方法org.apache.spark.sql.streaming.Trigger
-
A trigger policy that runs a query periodically based on an interval in processing time.
- processingTime() - 类 中的方法org.apache.spark.status.api.v1.streaming.BatchInfo
-
- ProcessingTimeTimeout() - 类 中的静态方法org.apache.spark.sql.streaming.GroupStateTimeout
-
Timeout based on processing time.
- processStreamByLine(String, InputStream, Function1<String, BoxedUnit>) - 类 中的静态方法org.apache.spark.util.Utils
-
Return and start a daemon thread that processes the content of the input stream line by line.
- producedAttributes() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- producedAttributes() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- producedAttributes() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- producedAttributes() - 类 中的方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- product() - 类 中的方法org.apache.spark.mllib.recommendation.Rating
-
- product(TypeTags.TypeTag<T>) - 类 中的静态方法org.apache.spark.sql.Encoders
-
An encoder for Scala's product type (tuples, case classes, etc).
- product(Numeric<B>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- productArity() - 类 中的静态方法org.apache.spark.Aggregator
-
- productArity() - 类 中的静态方法org.apache.spark.CleanAccum
-
- productArity() - 类 中的静态方法org.apache.spark.CleanBroadcast
-
- productArity() - 类 中的静态方法org.apache.spark.CleanCheckpoint
-
- productArity() - 类 中的静态方法org.apache.spark.CleanRDD
-
- productArity() - 类 中的静态方法org.apache.spark.CleanShuffle
-
- productArity() - 类 中的静态方法org.apache.spark.ExceptionFailure
-
- productArity() - 类 中的静态方法org.apache.spark.ExecutorLostFailure
-
- productArity() - 类 中的静态方法org.apache.spark.ExecutorRegistered
-
- productArity() - 类 中的静态方法org.apache.spark.ExecutorRemoved
-
- productArity() - 类 中的静态方法org.apache.spark.ExpireDeadHosts
-
- productArity() - 类 中的静态方法org.apache.spark.FetchFailed
-
- productArity() - 类 中的静态方法org.apache.spark.graphx.Edge
-
- productArity() - 类 中的静态方法org.apache.spark.ml.feature.Dot
-
- productArity() - 类 中的静态方法org.apache.spark.ml.feature.LabeledPoint
-
- productArity() - 类 中的静态方法org.apache.spark.ml.param.ParamPair
-
- productArity() - 类 中的静态方法org.apache.spark.mllib.feature.VocabWord
-
- productArity() - 类 中的静态方法org.apache.spark.mllib.linalg.distributed.IndexedRow
-
- productArity() - 类 中的静态方法org.apache.spark.mllib.linalg.distributed.MatrixEntry
-
- productArity() - 类 中的静态方法org.apache.spark.mllib.linalg.QRDecomposition
-
- productArity() - 类 中的静态方法org.apache.spark.mllib.linalg.SingularValueDecomposition
-
- productArity() - 类 中的静态方法org.apache.spark.mllib.recommendation.Rating
-
- productArity() - 类 中的静态方法org.apache.spark.mllib.regression.LabeledPoint
-
- productArity() - 类 中的静态方法org.apache.spark.mllib.stat.test.BinarySample
-
- productArity() - 类 中的静态方法org.apache.spark.mllib.tree.configuration.BoostingStrategy
-
- productArity() - 类 中的静态方法org.apache.spark.mllib.tree.model.Split
-
- productArity() - 类 中的静态方法org.apache.spark.Resubmitted
-
- productArity() - 类 中的静态方法org.apache.spark.rpc.netty.OnStart
-
- productArity() - 类 中的静态方法org.apache.spark.rpc.netty.OnStop
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.AccumulableInfo
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.AllJobsCancelled
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.AskPermissionToCommitOutput
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.BlacklistedExecutor
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.JobSucceeded
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.local.KillTask
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.local.ReviveOffers
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.local.StatusUpdate
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.local.StopExecutor
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.ResubmitFailedStages
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.RuntimePercentage
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerApplicationEnd
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerApplicationStart
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerBlockManagerAdded
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerBlockManagerRemoved
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerBlockUpdated
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerEnvironmentUpdate
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerExecutorAdded
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerExecutorBlacklisted
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerExecutorMetricsUpdate
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerExecutorRemoved
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerExecutorUnblacklisted
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerJobEnd
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerJobStart
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerLogStart
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerNodeBlacklisted
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerNodeUnblacklisted
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerSpeculativeTaskSubmitted
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerStageCompleted
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerStageSubmitted
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerTaskEnd
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerTaskGettingResult
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerTaskStart
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerUnpersistRDD
-
- productArity() - 类 中的静态方法org.apache.spark.scheduler.StopCoordinator
-
- productArity() - 类 中的静态方法org.apache.spark.sql.DatasetHolder
-
- productArity() - 类 中的静态方法org.apache.spark.sql.expressions.UserDefinedFunction
-
- productArity() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- productArity() - 类 中的静态方法org.apache.spark.sql.hive.execution.HiveScriptIOSchema
-
- productArity() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- productArity() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- productArity() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- productArity() - 类 中的静态方法org.apache.spark.sql.hive.RelationConversions
-
- productArity() - 类 中的静态方法org.apache.spark.sql.jdbc.JdbcType
-
- productArity() - 类 中的静态方法org.apache.spark.sql.jdbc.MySQLDialect
-
- productArity() - 类 中的静态方法org.apache.spark.sql.jdbc.OracleDialect
-
- productArity() - 类 中的静态方法org.apache.spark.sql.jdbc.TeradataDialect
-
- productArity() - 类 中的静态方法org.apache.spark.sql.sources.And
-
- productArity() - 类 中的静态方法org.apache.spark.sql.sources.EqualNullSafe
-
- productArity() - 类 中的静态方法org.apache.spark.sql.sources.EqualTo
-
- productArity() - 类 中的静态方法org.apache.spark.sql.sources.GreaterThan
-
- productArity() - 类 中的静态方法org.apache.spark.sql.sources.GreaterThanOrEqual
-
- productArity() - 类 中的静态方法org.apache.spark.sql.sources.In
-
- productArity() - 类 中的静态方法org.apache.spark.sql.sources.IsNotNull
-
- productArity() - 类 中的静态方法org.apache.spark.sql.sources.IsNull
-
- productArity() - 类 中的静态方法org.apache.spark.sql.sources.LessThan
-
- productArity() - 类 中的静态方法org.apache.spark.sql.sources.LessThanOrEqual
-
- productArity() - 类 中的静态方法org.apache.spark.sql.sources.Not
-
- productArity() - 类 中的静态方法org.apache.spark.sql.sources.Or
-
- productArity() - 类 中的静态方法org.apache.spark.sql.sources.StringContains
-
- productArity() - 类 中的静态方法org.apache.spark.sql.sources.StringEndsWith
-
- productArity() - 类 中的静态方法org.apache.spark.sql.sources.StringStartsWith
-
- productArity() - 类 中的静态方法org.apache.spark.sql.streaming.ProcessingTime
-
已过时。
- productArity() - 类 中的静态方法org.apache.spark.sql.types.ArrayType
-
- productArity() - 类 中的静态方法org.apache.spark.sql.types.CharType
-
- productArity() - 类 中的静态方法org.apache.spark.sql.types.DecimalType
-
- productArity() - 类 中的静态方法org.apache.spark.sql.types.MapType
-
- productArity() - 类 中的静态方法org.apache.spark.sql.types.ObjectType
-
- productArity() - 类 中的静态方法org.apache.spark.sql.types.StructField
-
- productArity() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- productArity() - 类 中的静态方法org.apache.spark.sql.types.VarcharType
-
- productArity() - 类 中的静态方法org.apache.spark.status.api.v1.ApplicationAttemptInfo
-
- productArity() - 类 中的静态方法org.apache.spark.status.api.v1.ApplicationInfo
-
- productArity() - 类 中的静态方法org.apache.spark.StopMapOutputTracker
-
- productArity() - 类 中的静态方法org.apache.spark.storage.BlockStatus
-
- productArity() - 类 中的静态方法org.apache.spark.storage.BlockUpdatedInfo
-
- productArity() - 类 中的静态方法org.apache.spark.storage.BroadcastBlockId
-
- productArity() - 类 中的静态方法org.apache.spark.storage.memory.DeserializedMemoryEntry
-
- productArity() - 类 中的静态方法org.apache.spark.storage.memory.SerializedMemoryEntry
-
- productArity() - 类 中的静态方法org.apache.spark.storage.RDDBlockId
-
- productArity() - 类 中的静态方法org.apache.spark.storage.ShuffleBlockId
-
- productArity() - 类 中的静态方法org.apache.spark.storage.ShuffleDataBlockId
-
- productArity() - 类 中的静态方法org.apache.spark.storage.ShuffleIndexBlockId
-
- productArity() - 类 中的静态方法org.apache.spark.storage.StreamBlockId
-
- productArity() - 类 中的静态方法org.apache.spark.storage.TaskResultBlockId
-
- productArity() - 类 中的静态方法org.apache.spark.streaming.Duration
-
- productArity() - 类 中的静态方法org.apache.spark.streaming.kinesis.DefaultCredentials
-
- productArity() - 类 中的静态方法org.apache.spark.streaming.scheduler.AllReceiverIds
-
- productArity() - 类 中的静态方法org.apache.spark.streaming.scheduler.BatchInfo
-
- productArity() - 类 中的静态方法org.apache.spark.streaming.scheduler.GetAllReceiverInfo
-
- productArity() - 类 中的静态方法org.apache.spark.streaming.scheduler.OutputOperationInfo
-
- productArity() - 类 中的静态方法org.apache.spark.streaming.scheduler.ReceiverInfo
-
- productArity() - 类 中的静态方法org.apache.spark.streaming.scheduler.StopAllReceivers
-
- productArity() - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerBatchCompleted
-
- productArity() - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerBatchStarted
-
- productArity() - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerBatchSubmitted
-
- productArity() - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerOutputOperationCompleted
-
- productArity() - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerOutputOperationStarted
-
- productArity() - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerReceiverError
-
- productArity() - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerReceiverStarted
-
- productArity() - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerReceiverStopped
-
- productArity() - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerStreamingStarted
-
- productArity() - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamInputInfo
-
- productArity() - 类 中的静态方法org.apache.spark.streaming.Time
-
- productArity() - 类 中的静态方法org.apache.spark.Success
-
- productArity() - 类 中的静态方法org.apache.spark.TaskCommitDenied
-
- productArity() - 类 中的静态方法org.apache.spark.TaskKilled
-
- productArity() - 类 中的静态方法org.apache.spark.TaskResultLost
-
- productArity() - 类 中的静态方法org.apache.spark.TaskSchedulerIsSet
-
- productArity() - 类 中的静态方法org.apache.spark.UnknownReason
-
- productArity() - 类 中的静态方法org.apache.spark.util.MethodIdentifier
-
- productArity() - 类 中的静态方法org.apache.spark.util.MutablePair
-
- productElement(int) - 类 中的静态方法org.apache.spark.Aggregator
-
- productElement(int) - 类 中的静态方法org.apache.spark.CleanAccum
-
- productElement(int) - 类 中的静态方法org.apache.spark.CleanBroadcast
-
- productElement(int) - 类 中的静态方法org.apache.spark.CleanCheckpoint
-
- productElement(int) - 类 中的静态方法org.apache.spark.CleanRDD
-
- productElement(int) - 类 中的静态方法org.apache.spark.CleanShuffle
-
- productElement(int) - 类 中的静态方法org.apache.spark.ExceptionFailure
-
- productElement(int) - 类 中的静态方法org.apache.spark.ExecutorLostFailure
-
- productElement(int) - 类 中的静态方法org.apache.spark.ExecutorRegistered
-
- productElement(int) - 类 中的静态方法org.apache.spark.ExecutorRemoved
-
- productElement(int) - 类 中的静态方法org.apache.spark.ExpireDeadHosts
-
- productElement(int) - 类 中的静态方法org.apache.spark.FetchFailed
-
- productElement(int) - 类 中的静态方法org.apache.spark.graphx.Edge
-
- productElement(int) - 类 中的静态方法org.apache.spark.ml.feature.Dot
-
- productElement(int) - 类 中的静态方法org.apache.spark.ml.feature.LabeledPoint
-
- productElement(int) - 类 中的静态方法org.apache.spark.ml.param.ParamPair
-
- productElement(int) - 类 中的静态方法org.apache.spark.mllib.feature.VocabWord
-
- productElement(int) - 类 中的静态方法org.apache.spark.mllib.linalg.distributed.IndexedRow
-
- productElement(int) - 类 中的静态方法org.apache.spark.mllib.linalg.distributed.MatrixEntry
-
- productElement(int) - 类 中的静态方法org.apache.spark.mllib.linalg.QRDecomposition
-
- productElement(int) - 类 中的静态方法org.apache.spark.mllib.linalg.SingularValueDecomposition
-
- productElement(int) - 类 中的静态方法org.apache.spark.mllib.recommendation.Rating
-
- productElement(int) - 类 中的静态方法org.apache.spark.mllib.regression.LabeledPoint
-
- productElement(int) - 类 中的静态方法org.apache.spark.mllib.stat.test.BinarySample
-
- productElement(int) - 类 中的静态方法org.apache.spark.mllib.tree.configuration.BoostingStrategy
-
- productElement(int) - 类 中的静态方法org.apache.spark.mllib.tree.model.Split
-
- productElement(int) - 类 中的静态方法org.apache.spark.Resubmitted
-
- productElement(int) - 类 中的静态方法org.apache.spark.rpc.netty.OnStart
-
- productElement(int) - 类 中的静态方法org.apache.spark.rpc.netty.OnStop
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.AccumulableInfo
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.AllJobsCancelled
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.AskPermissionToCommitOutput
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.BlacklistedExecutor
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.JobSucceeded
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.local.KillTask
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.local.ReviveOffers
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.local.StatusUpdate
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.local.StopExecutor
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.ResubmitFailedStages
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.RuntimePercentage
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerApplicationEnd
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerApplicationStart
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerBlockManagerAdded
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerBlockManagerRemoved
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerBlockUpdated
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerEnvironmentUpdate
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerExecutorAdded
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerExecutorBlacklisted
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerExecutorMetricsUpdate
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerExecutorRemoved
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerExecutorUnblacklisted
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerJobEnd
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerJobStart
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerLogStart
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerNodeBlacklisted
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerNodeUnblacklisted
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerSpeculativeTaskSubmitted
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerStageCompleted
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerStageSubmitted
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerTaskEnd
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerTaskGettingResult
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerTaskStart
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.SparkListenerUnpersistRDD
-
- productElement(int) - 类 中的静态方法org.apache.spark.scheduler.StopCoordinator
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.DatasetHolder
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.expressions.UserDefinedFunction
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.hive.execution.HiveScriptIOSchema
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.hive.RelationConversions
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.jdbc.JdbcType
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.jdbc.MySQLDialect
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.jdbc.OracleDialect
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.jdbc.TeradataDialect
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.sources.And
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.sources.EqualNullSafe
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.sources.EqualTo
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.sources.GreaterThan
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.sources.GreaterThanOrEqual
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.sources.In
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.sources.IsNotNull
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.sources.IsNull
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.sources.LessThan
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.sources.LessThanOrEqual
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.sources.Not
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.sources.Or
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.sources.StringContains
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.sources.StringEndsWith
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.sources.StringStartsWith
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.streaming.ProcessingTime
-
已过时。
- productElement(int) - 类 中的静态方法org.apache.spark.sql.types.ArrayType
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.types.CharType
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.types.DecimalType
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.types.MapType
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.types.ObjectType
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.types.StructField
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- productElement(int) - 类 中的静态方法org.apache.spark.sql.types.VarcharType
-
- productElement(int) - 类 中的静态方法org.apache.spark.status.api.v1.ApplicationAttemptInfo
-
- productElement(int) - 类 中的静态方法org.apache.spark.status.api.v1.ApplicationInfo
-
- productElement(int) - 类 中的静态方法org.apache.spark.StopMapOutputTracker
-
- productElement(int) - 类 中的静态方法org.apache.spark.storage.BlockStatus
-
- productElement(int) - 类 中的静态方法org.apache.spark.storage.BlockUpdatedInfo
-
- productElement(int) - 类 中的静态方法org.apache.spark.storage.BroadcastBlockId
-
- productElement(int) - 类 中的静态方法org.apache.spark.storage.memory.DeserializedMemoryEntry
-
- productElement(int) - 类 中的静态方法org.apache.spark.storage.memory.SerializedMemoryEntry
-
- productElement(int) - 类 中的静态方法org.apache.spark.storage.RDDBlockId
-
- productElement(int) - 类 中的静态方法org.apache.spark.storage.ShuffleBlockId
-
- productElement(int) - 类 中的静态方法org.apache.spark.storage.ShuffleDataBlockId
-
- productElement(int) - 类 中的静态方法org.apache.spark.storage.ShuffleIndexBlockId
-
- productElement(int) - 类 中的静态方法org.apache.spark.storage.StreamBlockId
-
- productElement(int) - 类 中的静态方法org.apache.spark.storage.TaskResultBlockId
-
- productElement(int) - 类 中的静态方法org.apache.spark.streaming.Duration
-
- productElement(int) - 类 中的静态方法org.apache.spark.streaming.kinesis.DefaultCredentials
-
- productElement(int) - 类 中的静态方法org.apache.spark.streaming.scheduler.AllReceiverIds
-
- productElement(int) - 类 中的静态方法org.apache.spark.streaming.scheduler.BatchInfo
-
- productElement(int) - 类 中的静态方法org.apache.spark.streaming.scheduler.GetAllReceiverInfo
-
- productElement(int) - 类 中的静态方法org.apache.spark.streaming.scheduler.OutputOperationInfo
-
- productElement(int) - 类 中的静态方法org.apache.spark.streaming.scheduler.ReceiverInfo
-
- productElement(int) - 类 中的静态方法org.apache.spark.streaming.scheduler.StopAllReceivers
-
- productElement(int) - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerBatchCompleted
-
- productElement(int) - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerBatchStarted
-
- productElement(int) - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerBatchSubmitted
-
- productElement(int) - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerOutputOperationCompleted
-
- productElement(int) - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerOutputOperationStarted
-
- productElement(int) - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerReceiverError
-
- productElement(int) - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerReceiverStarted
-
- productElement(int) - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerReceiverStopped
-
- productElement(int) - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerStreamingStarted
-
- productElement(int) - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamInputInfo
-
- productElement(int) - 类 中的静态方法org.apache.spark.streaming.Time
-
- productElement(int) - 类 中的静态方法org.apache.spark.Success
-
- productElement(int) - 类 中的静态方法org.apache.spark.TaskCommitDenied
-
- productElement(int) - 类 中的静态方法org.apache.spark.TaskKilled
-
- productElement(int) - 类 中的静态方法org.apache.spark.TaskResultLost
-
- productElement(int) - 类 中的静态方法org.apache.spark.TaskSchedulerIsSet
-
- productElement(int) - 类 中的静态方法org.apache.spark.UnknownReason
-
- productElement(int) - 类 中的静态方法org.apache.spark.util.MethodIdentifier
-
- productElement(int) - 类 中的静态方法org.apache.spark.util.MutablePair
-
- productFeatures() - 类 中的方法org.apache.spark.mllib.recommendation.MatrixFactorizationModel
-
- productIterator() - 类 中的静态方法org.apache.spark.Aggregator
-
- productIterator() - 类 中的静态方法org.apache.spark.CleanAccum
-
- productIterator() - 类 中的静态方法org.apache.spark.CleanBroadcast
-
- productIterator() - 类 中的静态方法org.apache.spark.CleanCheckpoint
-
- productIterator() - 类 中的静态方法org.apache.spark.CleanRDD
-
- productIterator() - 类 中的静态方法org.apache.spark.CleanShuffle
-
- productIterator() - 类 中的静态方法org.apache.spark.ExceptionFailure
-
- productIterator() - 类 中的静态方法org.apache.spark.ExecutorLostFailure
-
- productIterator() - 类 中的静态方法org.apache.spark.ExecutorRegistered
-
- productIterator() - 类 中的静态方法org.apache.spark.ExecutorRemoved
-
- productIterator() - 类 中的静态方法org.apache.spark.ExpireDeadHosts
-
- productIterator() - 类 中的静态方法org.apache.spark.FetchFailed
-
- productIterator() - 类 中的静态方法org.apache.spark.graphx.Edge
-
- productIterator() - 类 中的静态方法org.apache.spark.ml.feature.Dot
-
- productIterator() - 类 中的静态方法org.apache.spark.ml.feature.LabeledPoint
-
- productIterator() - 类 中的静态方法org.apache.spark.ml.param.ParamPair
-
- productIterator() - 类 中的静态方法org.apache.spark.mllib.feature.VocabWord
-
- productIterator() - 类 中的静态方法org.apache.spark.mllib.linalg.distributed.IndexedRow
-
- productIterator() - 类 中的静态方法org.apache.spark.mllib.linalg.distributed.MatrixEntry
-
- productIterator() - 类 中的静态方法org.apache.spark.mllib.linalg.QRDecomposition
-
- productIterator() - 类 中的静态方法org.apache.spark.mllib.linalg.SingularValueDecomposition
-
- productIterator() - 类 中的静态方法org.apache.spark.mllib.recommendation.Rating
-
- productIterator() - 类 中的静态方法org.apache.spark.mllib.regression.LabeledPoint
-
- productIterator() - 类 中的静态方法org.apache.spark.mllib.stat.test.BinarySample
-
- productIterator() - 类 中的静态方法org.apache.spark.mllib.tree.configuration.BoostingStrategy
-
- productIterator() - 类 中的静态方法org.apache.spark.mllib.tree.model.Split
-
- productIterator() - 类 中的静态方法org.apache.spark.Resubmitted
-
- productIterator() - 类 中的静态方法org.apache.spark.rpc.netty.OnStart
-
- productIterator() - 类 中的静态方法org.apache.spark.rpc.netty.OnStop
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.AccumulableInfo
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.AllJobsCancelled
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.AskPermissionToCommitOutput
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.BlacklistedExecutor
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.JobSucceeded
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.local.KillTask
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.local.ReviveOffers
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.local.StatusUpdate
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.local.StopExecutor
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.ResubmitFailedStages
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.RuntimePercentage
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerApplicationEnd
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerApplicationStart
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerBlockManagerAdded
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerBlockManagerRemoved
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerBlockUpdated
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerEnvironmentUpdate
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerExecutorAdded
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerExecutorBlacklisted
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerExecutorMetricsUpdate
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerExecutorRemoved
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerExecutorUnblacklisted
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerJobEnd
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerJobStart
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerLogStart
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerNodeBlacklisted
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerNodeUnblacklisted
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerSpeculativeTaskSubmitted
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerStageCompleted
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerStageSubmitted
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerTaskEnd
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerTaskGettingResult
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerTaskStart
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerUnpersistRDD
-
- productIterator() - 类 中的静态方法org.apache.spark.scheduler.StopCoordinator
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.DatasetHolder
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.expressions.UserDefinedFunction
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.hive.execution.HiveScriptIOSchema
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.hive.RelationConversions
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.jdbc.JdbcType
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.jdbc.MySQLDialect
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.jdbc.OracleDialect
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.jdbc.TeradataDialect
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.sources.And
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.sources.EqualNullSafe
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.sources.EqualTo
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.sources.GreaterThan
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.sources.GreaterThanOrEqual
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.sources.In
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.sources.IsNotNull
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.sources.IsNull
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.sources.LessThan
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.sources.LessThanOrEqual
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.sources.Not
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.sources.Or
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.sources.StringContains
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.sources.StringEndsWith
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.sources.StringStartsWith
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.streaming.ProcessingTime
-
已过时。
- productIterator() - 类 中的静态方法org.apache.spark.sql.types.ArrayType
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.types.CharType
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.types.DecimalType
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.types.MapType
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.types.ObjectType
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.types.StructField
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- productIterator() - 类 中的静态方法org.apache.spark.sql.types.VarcharType
-
- productIterator() - 类 中的静态方法org.apache.spark.status.api.v1.ApplicationAttemptInfo
-
- productIterator() - 类 中的静态方法org.apache.spark.status.api.v1.ApplicationInfo
-
- productIterator() - 类 中的静态方法org.apache.spark.StopMapOutputTracker
-
- productIterator() - 类 中的静态方法org.apache.spark.storage.BlockStatus
-
- productIterator() - 类 中的静态方法org.apache.spark.storage.BlockUpdatedInfo
-
- productIterator() - 类 中的静态方法org.apache.spark.storage.BroadcastBlockId
-
- productIterator() - 类 中的静态方法org.apache.spark.storage.memory.DeserializedMemoryEntry
-
- productIterator() - 类 中的静态方法org.apache.spark.storage.memory.SerializedMemoryEntry
-
- productIterator() - 类 中的静态方法org.apache.spark.storage.RDDBlockId
-
- productIterator() - 类 中的静态方法org.apache.spark.storage.ShuffleBlockId
-
- productIterator() - 类 中的静态方法org.apache.spark.storage.ShuffleDataBlockId
-
- productIterator() - 类 中的静态方法org.apache.spark.storage.ShuffleIndexBlockId
-
- productIterator() - 类 中的静态方法org.apache.spark.storage.StreamBlockId
-
- productIterator() - 类 中的静态方法org.apache.spark.storage.TaskResultBlockId
-
- productIterator() - 类 中的静态方法org.apache.spark.streaming.Duration
-
- productIterator() - 类 中的静态方法org.apache.spark.streaming.kinesis.DefaultCredentials
-
- productIterator() - 类 中的静态方法org.apache.spark.streaming.scheduler.AllReceiverIds
-
- productIterator() - 类 中的静态方法org.apache.spark.streaming.scheduler.BatchInfo
-
- productIterator() - 类 中的静态方法org.apache.spark.streaming.scheduler.GetAllReceiverInfo
-
- productIterator() - 类 中的静态方法org.apache.spark.streaming.scheduler.OutputOperationInfo
-
- productIterator() - 类 中的静态方法org.apache.spark.streaming.scheduler.ReceiverInfo
-
- productIterator() - 类 中的静态方法org.apache.spark.streaming.scheduler.StopAllReceivers
-
- productIterator() - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerBatchCompleted
-
- productIterator() - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerBatchStarted
-
- productIterator() - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerBatchSubmitted
-
- productIterator() - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerOutputOperationCompleted
-
- productIterator() - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerOutputOperationStarted
-
- productIterator() - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerReceiverError
-
- productIterator() - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerReceiverStarted
-
- productIterator() - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerReceiverStopped
-
- productIterator() - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerStreamingStarted
-
- productIterator() - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamInputInfo
-
- productIterator() - 类 中的静态方法org.apache.spark.streaming.Time
-
- productIterator() - 类 中的静态方法org.apache.spark.Success
-
- productIterator() - 类 中的静态方法org.apache.spark.TaskCommitDenied
-
- productIterator() - 类 中的静态方法org.apache.spark.TaskKilled
-
- productIterator() - 类 中的静态方法org.apache.spark.TaskResultLost
-
- productIterator() - 类 中的静态方法org.apache.spark.TaskSchedulerIsSet
-
- productIterator() - 类 中的静态方法org.apache.spark.UnknownReason
-
- productIterator() - 类 中的静态方法org.apache.spark.util.MethodIdentifier
-
- productIterator() - 类 中的静态方法org.apache.spark.util.MutablePair
-
- productPrefix() - 类 中的静态方法org.apache.spark.Aggregator
-
- productPrefix() - 类 中的静态方法org.apache.spark.CleanAccum
-
- productPrefix() - 类 中的静态方法org.apache.spark.CleanBroadcast
-
- productPrefix() - 类 中的静态方法org.apache.spark.CleanCheckpoint
-
- productPrefix() - 类 中的静态方法org.apache.spark.CleanRDD
-
- productPrefix() - 类 中的静态方法org.apache.spark.CleanShuffle
-
- productPrefix() - 类 中的静态方法org.apache.spark.ExceptionFailure
-
- productPrefix() - 类 中的静态方法org.apache.spark.ExecutorLostFailure
-
- productPrefix() - 类 中的静态方法org.apache.spark.ExecutorRegistered
-
- productPrefix() - 类 中的静态方法org.apache.spark.ExecutorRemoved
-
- productPrefix() - 类 中的静态方法org.apache.spark.ExpireDeadHosts
-
- productPrefix() - 类 中的静态方法org.apache.spark.FetchFailed
-
- productPrefix() - 类 中的静态方法org.apache.spark.graphx.Edge
-
- productPrefix() - 类 中的静态方法org.apache.spark.ml.feature.Dot
-
- productPrefix() - 类 中的静态方法org.apache.spark.ml.feature.LabeledPoint
-
- productPrefix() - 类 中的静态方法org.apache.spark.ml.param.ParamPair
-
- productPrefix() - 类 中的静态方法org.apache.spark.mllib.feature.VocabWord
-
- productPrefix() - 类 中的静态方法org.apache.spark.mllib.linalg.distributed.IndexedRow
-
- productPrefix() - 类 中的静态方法org.apache.spark.mllib.linalg.distributed.MatrixEntry
-
- productPrefix() - 类 中的静态方法org.apache.spark.mllib.linalg.QRDecomposition
-
- productPrefix() - 类 中的静态方法org.apache.spark.mllib.linalg.SingularValueDecomposition
-
- productPrefix() - 类 中的静态方法org.apache.spark.mllib.recommendation.Rating
-
- productPrefix() - 类 中的静态方法org.apache.spark.mllib.regression.LabeledPoint
-
- productPrefix() - 类 中的静态方法org.apache.spark.mllib.stat.test.BinarySample
-
- productPrefix() - 类 中的静态方法org.apache.spark.mllib.tree.configuration.BoostingStrategy
-
- productPrefix() - 类 中的静态方法org.apache.spark.mllib.tree.model.Split
-
- productPrefix() - 类 中的静态方法org.apache.spark.Resubmitted
-
- productPrefix() - 类 中的静态方法org.apache.spark.rpc.netty.OnStart
-
- productPrefix() - 类 中的静态方法org.apache.spark.rpc.netty.OnStop
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.AccumulableInfo
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.AllJobsCancelled
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.AskPermissionToCommitOutput
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.BlacklistedExecutor
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.JobSucceeded
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.local.KillTask
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.local.ReviveOffers
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.local.StatusUpdate
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.local.StopExecutor
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.ResubmitFailedStages
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.RuntimePercentage
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerApplicationEnd
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerApplicationStart
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerBlockManagerAdded
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerBlockManagerRemoved
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerBlockUpdated
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerEnvironmentUpdate
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerExecutorAdded
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerExecutorBlacklisted
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerExecutorMetricsUpdate
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerExecutorRemoved
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerExecutorUnblacklisted
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerJobEnd
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerJobStart
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerLogStart
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerNodeBlacklisted
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerNodeUnblacklisted
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerSpeculativeTaskSubmitted
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerStageCompleted
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerStageSubmitted
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerTaskEnd
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerTaskGettingResult
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerTaskStart
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.SparkListenerUnpersistRDD
-
- productPrefix() - 类 中的静态方法org.apache.spark.scheduler.StopCoordinator
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.DatasetHolder
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.expressions.UserDefinedFunction
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.hive.execution.HiveScriptIOSchema
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.hive.RelationConversions
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.jdbc.JdbcType
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.jdbc.MySQLDialect
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.jdbc.OracleDialect
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.jdbc.TeradataDialect
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.sources.And
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.sources.EqualNullSafe
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.sources.EqualTo
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.sources.GreaterThan
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.sources.GreaterThanOrEqual
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.sources.In
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.sources.IsNotNull
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.sources.IsNull
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.sources.LessThan
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.sources.LessThanOrEqual
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.sources.Not
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.sources.Or
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.sources.StringContains
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.sources.StringEndsWith
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.sources.StringStartsWith
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.streaming.ProcessingTime
-
已过时。
- productPrefix() - 类 中的静态方法org.apache.spark.sql.types.ArrayType
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.types.CharType
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.types.DecimalType
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.types.MapType
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.types.ObjectType
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.types.StructField
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- productPrefix() - 类 中的静态方法org.apache.spark.sql.types.VarcharType
-
- productPrefix() - 类 中的静态方法org.apache.spark.status.api.v1.ApplicationAttemptInfo
-
- productPrefix() - 类 中的静态方法org.apache.spark.status.api.v1.ApplicationInfo
-
- productPrefix() - 类 中的静态方法org.apache.spark.StopMapOutputTracker
-
- productPrefix() - 类 中的静态方法org.apache.spark.storage.BlockStatus
-
- productPrefix() - 类 中的静态方法org.apache.spark.storage.BlockUpdatedInfo
-
- productPrefix() - 类 中的静态方法org.apache.spark.storage.BroadcastBlockId
-
- productPrefix() - 类 中的静态方法org.apache.spark.storage.memory.DeserializedMemoryEntry
-
- productPrefix() - 类 中的静态方法org.apache.spark.storage.memory.SerializedMemoryEntry
-
- productPrefix() - 类 中的静态方法org.apache.spark.storage.RDDBlockId
-
- productPrefix() - 类 中的静态方法org.apache.spark.storage.ShuffleBlockId
-
- productPrefix() - 类 中的静态方法org.apache.spark.storage.ShuffleDataBlockId
-
- productPrefix() - 类 中的静态方法org.apache.spark.storage.ShuffleIndexBlockId
-
- productPrefix() - 类 中的静态方法org.apache.spark.storage.StreamBlockId
-
- productPrefix() - 类 中的静态方法org.apache.spark.storage.TaskResultBlockId
-
- productPrefix() - 类 中的静态方法org.apache.spark.streaming.Duration
-
- productPrefix() - 类 中的静态方法org.apache.spark.streaming.kinesis.DefaultCredentials
-
- productPrefix() - 类 中的静态方法org.apache.spark.streaming.scheduler.AllReceiverIds
-
- productPrefix() - 类 中的静态方法org.apache.spark.streaming.scheduler.BatchInfo
-
- productPrefix() - 类 中的静态方法org.apache.spark.streaming.scheduler.GetAllReceiverInfo
-
- productPrefix() - 类 中的静态方法org.apache.spark.streaming.scheduler.OutputOperationInfo
-
- productPrefix() - 类 中的静态方法org.apache.spark.streaming.scheduler.ReceiverInfo
-
- productPrefix() - 类 中的静态方法org.apache.spark.streaming.scheduler.StopAllReceivers
-
- productPrefix() - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerBatchCompleted
-
- productPrefix() - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerBatchStarted
-
- productPrefix() - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerBatchSubmitted
-
- productPrefix() - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerOutputOperationCompleted
-
- productPrefix() - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerOutputOperationStarted
-
- productPrefix() - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerReceiverError
-
- productPrefix() - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerReceiverStarted
-
- productPrefix() - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerReceiverStopped
-
- productPrefix() - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamingListenerStreamingStarted
-
- productPrefix() - 类 中的静态方法org.apache.spark.streaming.scheduler.StreamInputInfo
-
- productPrefix() - 类 中的静态方法org.apache.spark.streaming.Time
-
- productPrefix() - 类 中的静态方法org.apache.spark.Success
-
- productPrefix() - 类 中的静态方法org.apache.spark.TaskCommitDenied
-
- productPrefix() - 类 中的静态方法org.apache.spark.TaskKilled
-
- productPrefix() - 类 中的静态方法org.apache.spark.TaskResultLost
-
- productPrefix() - 类 中的静态方法org.apache.spark.TaskSchedulerIsSet
-
- productPrefix() - 类 中的静态方法org.apache.spark.UnknownReason
-
- productPrefix() - 类 中的静态方法org.apache.spark.util.MethodIdentifier
-
- productPrefix() - 类 中的静态方法org.apache.spark.util.MutablePair
-
- progress() - 类 中的方法org.apache.spark.sql.streaming.StreamingQueryListener.QueryProgressEvent
-
- project(double) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.Binomial$
-
- project(double) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression.Gaussian$
-
- properties() - 类 中的方法org.apache.spark.scheduler.SparkListenerJobStart
-
- properties() - 类 中的方法org.apache.spark.scheduler.SparkListenerStageSubmitted
-
- propertiesFromJson(JsonAST.JValue) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- propertiesToJson(Properties) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- provider() - 类 中的静态方法org.apache.spark.streaming.kinesis.DefaultCredentials
-
- proxyBase() - 类 中的方法org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.AddWebUIFilter
-
- pruneColumns(StructType) - 接口 中的方法org.apache.spark.sql.sources.v2.reader.SupportsPushDownRequiredColumns
-
Applies column pruning w.r.t. the given requiredSchema.
- PrunedFilteredScan - org.apache.spark.sql.sources中的接口
-
A BaseRelation that can eliminate unneeded columns and filter using selected
predicates before producing an RDD containing all matching tuples as Row objects.
- PrunedScan - org.apache.spark.sql.sources中的接口
-
A BaseRelation that can eliminate unneeded columns before producing an RDD
containing all of its tuples as Row objects.
- Pseudorandom - org.apache.spark.util.random中的接口
-
:: DeveloperApi ::
A class with pseudorandom behavior.
- pushCatalystFilters(Expression[]) - 接口 中的方法org.apache.spark.sql.sources.v2.reader.SupportsPushDownCatalystFilters
-
Pushes down filters, and returns filters that need to be evaluated after scanning.
- pushedCatalystFilters() - 接口 中的方法org.apache.spark.sql.sources.v2.reader.SupportsPushDownCatalystFilters
-
- pushedFilters() - 接口 中的方法org.apache.spark.sql.sources.v2.reader.SupportsPushDownFilters
-
- pushFilters(Filter[]) - 接口 中的方法org.apache.spark.sql.sources.v2.reader.SupportsPushDownFilters
-
Pushes down filters, and returns filters that need to be evaluated after scanning.
- put(ParamPair<?>...) - 类 中的方法org.apache.spark.ml.param.ParamMap
-
Puts a list of param pairs (overwrites if the input params exists).
- put(Param<T>, T) - 类 中的方法org.apache.spark.ml.param.ParamMap
-
Puts a (param, value) pair (overwrites if the input param exists).
- put(Seq<ParamPair<?>>) - 类 中的方法org.apache.spark.ml.param.ParamMap
-
Puts a list of param pairs (overwrites if the input params exists).
- put(Object) - 类 中的方法org.apache.spark.util.sketch.BloomFilter
-
Puts an item into this BloomFilter.
- putBinary(byte[]) - 类 中的方法org.apache.spark.util.sketch.BloomFilter
-
- putBoolean(String, boolean) - 类 中的方法org.apache.spark.sql.types.MetadataBuilder
-
Puts a Boolean.
- putBooleanArray(String, boolean[]) - 类 中的方法org.apache.spark.sql.types.MetadataBuilder
-
Puts a Boolean array.
- putDouble(String, double) - 类 中的方法org.apache.spark.sql.types.MetadataBuilder
-
Puts a Double.
- putDoubleArray(String, double[]) - 类 中的方法org.apache.spark.sql.types.MetadataBuilder
-
Puts a Double array.
- putLong(String, long) - 类 中的方法org.apache.spark.sql.types.MetadataBuilder
-
Puts a Long.
- putLong(long) - 类 中的方法org.apache.spark.util.sketch.BloomFilter
-
- putLongArray(String, long[]) - 类 中的方法org.apache.spark.sql.types.MetadataBuilder
-
Puts a Long array.
- putMetadata(String, Metadata) - 类 中的方法org.apache.spark.sql.types.MetadataBuilder
-
- putMetadataArray(String, Metadata[]) - 类 中的方法org.apache.spark.sql.types.MetadataBuilder
-
- putNull(String) - 类 中的方法org.apache.spark.sql.types.MetadataBuilder
-
Puts a null.
- putString(String, String) - 类 中的方法org.apache.spark.sql.types.MetadataBuilder
-
Puts a String.
- putString(String) - 类 中的方法org.apache.spark.util.sketch.BloomFilter
-
- putStringArray(String, String[]) - 类 中的方法org.apache.spark.sql.types.MetadataBuilder
-
Puts a String array.
- pValue() - 类 中的方法org.apache.spark.mllib.stat.test.ChiSqTestResult
-
- pValue() - 类 中的方法org.apache.spark.mllib.stat.test.KolmogorovSmirnovTestResult
-
- pValue() - 接口 中的方法org.apache.spark.mllib.stat.test.TestResult
-
The probability of obtaining a test statistic result at least as extreme as the one that was
actually observed, assuming that the null hypothesis is true.
- pValues() - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegressionTrainingSummary
-
Two-sided p-value of estimated coefficients and intercept.
- pValues() - 类 中的方法org.apache.spark.ml.regression.LinearRegressionSummary
-
Two-sided p-value of estimated coefficients and intercept.
- pyUDT() - 类 中的方法org.apache.spark.mllib.linalg.VectorUDT
-
- R() - 类 中的方法org.apache.spark.mllib.linalg.QRDecomposition
-
- r2() - 类 中的方法org.apache.spark.ml.regression.LinearRegressionSummary
-
Returns R^2^, the coefficient of determination.
- r2() - 类 中的方法org.apache.spark.mllib.evaluation.RegressionMetrics
-
Returns R^2^, the unadjusted coefficient of determination.
- r2adj() - 类 中的方法org.apache.spark.ml.regression.LinearRegressionSummary
-
Returns Adjusted R^2^, the adjusted coefficient of determination.
- RACK_LOCAL() - 类 中的静态方法org.apache.spark.scheduler.TaskLocality
-
- radians(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Converts an angle measured in degrees to an approximately equivalent angle measured in radians.
- radians(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Converts an angle measured in degrees to an approximately equivalent angle measured in radians.
- rand(int, int, Random) - 类 中的静态方法org.apache.spark.ml.linalg.DenseMatrix
-
Generate a DenseMatrix consisting of i.i.d.
- rand(int, int, Random) - 类 中的静态方法org.apache.spark.ml.linalg.Matrices
-
Generate a DenseMatrix consisting of i.i.d.
- rand(int, int, Random) - 类 中的静态方法org.apache.spark.mllib.linalg.DenseMatrix
-
Generate a DenseMatrix consisting of i.i.d.
- rand(int, int, Random) - 类 中的静态方法org.apache.spark.mllib.linalg.Matrices
-
Generate a DenseMatrix consisting of i.i.d.
- rand(long) - 类 中的静态方法org.apache.spark.sql.functions
-
Generate a random column with independent and identically distributed (i.i.d.) samples
from U[0.0, 1.0].
- rand() - 类 中的静态方法org.apache.spark.sql.functions
-
Generate a random column with independent and identically distributed (i.i.d.) samples
from U[0.0, 1.0].
- randn(int, int, Random) - 类 中的静态方法org.apache.spark.ml.linalg.DenseMatrix
-
Generate a DenseMatrix consisting of i.i.d.
- randn(int, int, Random) - 类 中的静态方法org.apache.spark.ml.linalg.Matrices
-
Generate a DenseMatrix consisting of i.i.d.
- randn(int, int, Random) - 类 中的静态方法org.apache.spark.mllib.linalg.DenseMatrix
-
Generate a DenseMatrix consisting of i.i.d.
- randn(int, int, Random) - 类 中的静态方法org.apache.spark.mllib.linalg.Matrices
-
Generate a DenseMatrix consisting of i.i.d.
- randn(long) - 类 中的静态方法org.apache.spark.sql.functions
-
Generate a column with independent and identically distributed (i.i.d.) samples from
the standard normal distribution.
- randn() - 类 中的静态方法org.apache.spark.sql.functions
-
Generate a column with independent and identically distributed (i.i.d.) samples from
the standard normal distribution.
- random() - 类 中的方法org.apache.spark.ml.image.SamplePathFilter
-
- RANDOM() - 类 中的静态方法org.apache.spark.mllib.clustering.KMeans
-
- random() - 类 中的静态方法org.apache.spark.util.Utils
-
- RandomBlockReplicationPolicy - org.apache.spark.storage中的类
-
- RandomBlockReplicationPolicy() - 类 的构造器org.apache.spark.storage.RandomBlockReplicationPolicy
-
- RandomDataGenerator<T> - org.apache.spark.mllib.random中的接口
-
:: DeveloperApi ::
Trait for random data generators that generate i.i.d. data.
- RandomForest - org.apache.spark.ml.tree.impl中的类
-
ALGORITHM
This is a sketch of the algorithm to help new developers.
- RandomForest() - 类 的构造器org.apache.spark.ml.tree.impl.RandomForest
-
- RandomForest - org.apache.spark.mllib.tree中的类
-
A class that implements a
Random Forest
learning algorithm for classification and regression.
- RandomForest(Strategy, int, String, int) - 类 的构造器org.apache.spark.mllib.tree.RandomForest
-
- RandomForestClassificationModel - org.apache.spark.ml.classification中的类
-
- RandomForestClassifier - org.apache.spark.ml.classification中的类
-
- RandomForestClassifier(String) - 类 的构造器org.apache.spark.ml.classification.RandomForestClassifier
-
- RandomForestClassifier() - 类 的构造器org.apache.spark.ml.classification.RandomForestClassifier
-
- RandomForestModel - org.apache.spark.mllib.tree.model中的类
-
Represents a random forest model.
- RandomForestModel(Enumeration.Value, DecisionTreeModel[]) - 类 的构造器org.apache.spark.mllib.tree.model.RandomForestModel
-
- RandomForestRegressionModel - org.apache.spark.ml.regression中的类
-
- RandomForestRegressor - org.apache.spark.ml.regression中的类
-
- RandomForestRegressor(String) - 类 的构造器org.apache.spark.ml.regression.RandomForestRegressor
-
- RandomForestRegressor() - 类 的构造器org.apache.spark.ml.regression.RandomForestRegressor
-
- randomize(TraversableOnce<T>, ClassTag<T>) - 类 中的静态方法org.apache.spark.util.Utils
-
Shuffle the elements of a collection into a random order, returning the
result in a new collection.
- randomizeInPlace(Object, Random) - 类 中的静态方法org.apache.spark.util.Utils
-
Shuffle the elements of an array into a random order, modifying the
original array.
- randomJavaRDD(JavaSparkContext, RandomDataGenerator<T>, long, int, long) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
:: DeveloperApi ::
Generates an RDD comprised of i.i.d.
- randomJavaRDD(JavaSparkContext, RandomDataGenerator<T>, long, int) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
:: DeveloperApi ::
RandomRDDs.randomJavaRDD with the default seed.
- randomJavaRDD(JavaSparkContext, RandomDataGenerator<T>, long) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
:: DeveloperApi ::
RandomRDDs.randomJavaRDD with the default seed & numPartitions
- randomJavaVectorRDD(JavaSparkContext, RandomDataGenerator<Object>, long, int, int, long) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
:: DeveloperApi ::
Java-friendly version of RandomRDDs.randomVectorRDD.
- randomJavaVectorRDD(JavaSparkContext, RandomDataGenerator<Object>, long, int, int) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
:: DeveloperApi ::
RandomRDDs.randomJavaVectorRDD with the default seed.
- randomJavaVectorRDD(JavaSparkContext, RandomDataGenerator<Object>, long, int) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
:: DeveloperApi ::
RandomRDDs.randomJavaVectorRDD with the default number of partitions and the default seed.
- randomRDD(SparkContext, RandomDataGenerator<T>, long, int, long, ClassTag<T>) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
:: DeveloperApi ::
Generates an RDD comprised of i.i.d.
- RandomRDDs - org.apache.spark.mllib.random中的类
-
Generator methods for creating RDDs comprised of i.i.d.
- RandomRDDs() - 类 的构造器org.apache.spark.mllib.random.RandomRDDs
-
- RandomSampler<T,U> - org.apache.spark.util.random中的接口
-
:: DeveloperApi ::
A pseudorandom sampler.
- randomSplit(double[]) - 类 中的方法org.apache.spark.api.java.JavaRDD
-
Randomly splits this RDD with the provided weights.
- randomSplit(double[], long) - 类 中的方法org.apache.spark.api.java.JavaRDD
-
Randomly splits this RDD with the provided weights.
- randomSplit(double[], long) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- randomSplit(double[], long) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- randomSplit(double[], long) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- randomSplit(double[], long) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- randomSplit(double[], long) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- randomSplit(double[], long) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- randomSplit(double[], long) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- randomSplit(double[], long) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- randomSplit(double[], long) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- randomSplit(double[], long) - 类 中的方法org.apache.spark.rdd.RDD
-
Randomly splits this RDD with the provided weights.
- randomSplit(double[], long) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- randomSplit(double[], long) - 类 中的方法org.apache.spark.sql.Dataset
-
Randomly splits this Dataset with the provided weights.
- randomSplit(double[]) - 类 中的方法org.apache.spark.sql.Dataset
-
Randomly splits this Dataset with the provided weights.
- randomSplit$default$2() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- randomSplit$default$2() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- randomSplit$default$2() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- randomSplit$default$2() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- randomSplit$default$2() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- randomSplit$default$2() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- randomSplit$default$2() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- randomSplit$default$2() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- randomSplit$default$2() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- randomSplit$default$2() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- randomSplitAsList(double[], long) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a Java list that contains randomly split Dataset with the provided weights.
- randomVectorRDD(SparkContext, RandomDataGenerator<Object>, long, int, int, long) - 类 中的静态方法org.apache.spark.mllib.random.RandomRDDs
-
:: DeveloperApi ::
Generates an RDD[Vector] with vectors containing i.i.d.
- RandomVertexCut$() - 类 的构造器org.apache.spark.graphx.PartitionStrategy.RandomVertexCut$
-
- range(long, long, long, int) - 类 中的方法org.apache.spark.SparkContext
-
Creates a new RDD[Long] containing elements from start to end(exclusive), increased by
step every element.
- range(long) - 类 中的方法org.apache.spark.sql.SparkSession
-
:: Experimental ::
Creates a
Dataset with a single
LongType column named
id, containing elements
in a range from 0 to
end (exclusive) with step value 1.
- range(long, long) - 类 中的方法org.apache.spark.sql.SparkSession
-
:: Experimental ::
Creates a
Dataset with a single
LongType column named
id, containing elements
in a range from
start to
end (exclusive) with step value 1.
- range(long, long, long) - 类 中的方法org.apache.spark.sql.SparkSession
-
:: Experimental ::
Creates a
Dataset with a single
LongType column named
id, containing elements
in a range from
start to
end (exclusive) with a step value.
- range(long, long, long, int) - 类 中的方法org.apache.spark.sql.SparkSession
-
:: Experimental ::
Creates a
Dataset with a single
LongType column named
id, containing elements
in a range from
start to
end (exclusive) with a step value, with partition number
specified.
- range(long) - 类 中的方法org.apache.spark.sql.SQLContext
-
- range(long, long) - 类 中的方法org.apache.spark.sql.SQLContext
-
- range(long, long, long) - 类 中的方法org.apache.spark.sql.SQLContext
-
- range(long, long, long, int) - 类 中的方法org.apache.spark.sql.SQLContext
-
- rangeBetween(long, long) - 类 中的静态方法org.apache.spark.sql.expressions.Window
-
Creates a
WindowSpec with the frame boundaries defined,
from
start (inclusive) to
end (inclusive).
- rangeBetween(Column, Column) - 类 中的静态方法org.apache.spark.sql.expressions.Window
-
Creates a
WindowSpec with the frame boundaries defined,
from
start (inclusive) to
end (inclusive).
- rangeBetween(long, long) - 类 中的方法org.apache.spark.sql.expressions.WindowSpec
-
Defines the frame boundaries, from start (inclusive) to end (inclusive).
- rangeBetween(Column, Column) - 类 中的方法org.apache.spark.sql.expressions.WindowSpec
-
Defines the frame boundaries, from start (inclusive) to end (inclusive).
- RangeDependency<T> - org.apache.spark中的类
-
:: DeveloperApi ::
Represents a one-to-one dependency between ranges of partitions in the parent and child RDDs.
- RangeDependency(RDD<T>, int, int, int) - 类 的构造器org.apache.spark.RangeDependency
-
- RangePartitioner<K,V> - org.apache.spark中的类
-
A
Partitioner that partitions sortable records by range into roughly
equal ranges.
- RangePartitioner(int, RDD<? extends Product2<K, V>>, boolean, int, Ordering<K>, ClassTag<K>) - 类 的构造器org.apache.spark.RangePartitioner
-
- RangePartitioner(int, RDD<? extends Product2<K, V>>, boolean, Ordering<K>, ClassTag<K>) - 类 的构造器org.apache.spark.RangePartitioner
-
- rank() - 类 中的方法org.apache.spark.graphx.lib.SVDPlusPlus.Conf
-
- rank() - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- rank() - 类 中的方法org.apache.spark.ml.recommendation.ALSModel
-
- rank() - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegressionSummary
-
The numeric rank of the fitted linear model.
- rank() - 类 中的方法org.apache.spark.mllib.recommendation.MatrixFactorizationModel
-
- rank() - 类 中的静态方法org.apache.spark.sql.functions
-
Window function: returns the rank of rows within a window partition.
- RankingMetrics<T> - org.apache.spark.mllib.evaluation中的类
-
Evaluator for ranking algorithms.
- RankingMetrics(RDD<Tuple2<Object, Object>>, ClassTag<T>) - 类 的构造器org.apache.spark.mllib.evaluation.RankingMetrics
-
- Rating(ID, ID, float) - 类 的构造器org.apache.spark.ml.recommendation.ALS.Rating
-
- rating() - 类 中的方法org.apache.spark.ml.recommendation.ALS.Rating
-
- Rating - org.apache.spark.mllib.recommendation中的类
-
A more compact class to represent a rating than Tuple3[Int, Int, Double].
- Rating(int, int, double) - 类 的构造器org.apache.spark.mllib.recommendation.Rating
-
- rating() - 类 中的方法org.apache.spark.mllib.recommendation.Rating
-
- Rating$() - 类 的构造器org.apache.spark.ml.recommendation.ALS.Rating$
-
- RatingBlock$() - 类 的构造器org.apache.spark.ml.recommendation.ALS.RatingBlock$
-
- ratingCol() - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- ratioParam() - 类 中的静态方法org.apache.spark.ml.image.SamplePathFilter
-
- rawPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- rawPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- rawPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- rawPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- rawPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- rawPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- rawPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- rawPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- rawPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- rawPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- rawPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- rawPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- rawPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- rawPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- rawPredictionCol() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- rawPredictionCol() - 类 中的静态方法org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
-
- rawPredictionCol() - 接口 中的方法org.apache.spark.ml.param.shared.HasRawPredictionCol
-
Param for raw prediction (a.k.a. confidence) column name.
- rawSocketStream(String, int, StorageLevel) - 类 中的方法org.apache.spark.streaming.api.java.JavaStreamingContext
-
Create an input stream from network source hostname:port, where data is received
as serialized blocks (serialized using the Spark's serializer) that can be directly
pushed into the block manager without deserializing them.
- rawSocketStream(String, int) - 类 中的方法org.apache.spark.streaming.api.java.JavaStreamingContext
-
Create an input stream from network source hostname:port, where data is received
as serialized blocks (serialized using the Spark's serializer) that can be directly
pushed into the block manager without deserializing them.
- rawSocketStream(String, int, StorageLevel, ClassTag<T>) - 类 中的方法org.apache.spark.streaming.StreamingContext
-
Create an input stream from network source hostname:port, where data is received
as serialized blocks (serialized using the Spark's serializer) that can be directly
pushed into the block manager without deserializing them.
- RawTextHelper - org.apache.spark.streaming.util中的类
-
- RawTextHelper() - 类 的构造器org.apache.spark.streaming.util.RawTextHelper
-
- RawTextSender - org.apache.spark.streaming.util中的类
-
A helper program that sends blocks of Kryo-serialized text strings out on a socket at a
specified rate.
- RawTextSender() - 类 的构造器org.apache.spark.streaming.util.RawTextSender
-
- RBackendAuthHandler - org.apache.spark.api.r中的类
-
Authentication handler for connections from the R process.
- RBackendAuthHandler(String) - 类 的构造器org.apache.spark.api.r.RBackendAuthHandler
-
- rdd() - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
- rdd() - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
- rdd() - 类 中的方法org.apache.spark.api.java.JavaRDD
-
- rdd() - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
- RDD() - 类 中的静态方法org.apache.spark.api.r.RRunnerModes
-
- rdd() - 类 中的方法org.apache.spark.Dependency
-
- rdd() - 类 中的方法org.apache.spark.NarrowDependency
-
- RDD<T> - org.apache.spark.rdd中的类
-
A Resilient Distributed Dataset (RDD), the basic abstraction in Spark.
- RDD(SparkContext, Seq<Dependency<?>>, ClassTag<T>) - 类 的构造器org.apache.spark.rdd.RDD
-
- RDD(RDD<?>, ClassTag<T>) - 类 的构造器org.apache.spark.rdd.RDD
-
Construct an RDD with just a one-to-one dependency on one parent
- rdd() - 类 中的方法org.apache.spark.ShuffleDependency
-
- rdd() - 类 中的方法org.apache.spark.sql.Dataset
-
Represents the content of the Dataset as an RDD of T.
- RDD() - 类 中的静态方法org.apache.spark.storage.BlockId
-
- RDDBlockId - org.apache.spark.storage中的类
-
- RDDBlockId(int, int) - 类 的构造器org.apache.spark.storage.RDDBlockId
-
- rddBlocks() - 类 中的方法org.apache.spark.status.api.v1.ExecutorSummary
-
- rddBlocks() - 类 中的方法org.apache.spark.status.LiveExecutor
-
- rddBlocks() - 类 中的方法org.apache.spark.storage.StorageStatus
-
已过时。
Return the RDD blocks stored in this block manager.
- rddBlocksById(int) - 类 中的方法org.apache.spark.storage.StorageStatus
-
已过时。
Return the blocks that belong to the given RDD stored in this block manager.
- RDDDataDistribution - org.apache.spark.status.api.v1中的类
-
- RDDFunctions<T> - org.apache.spark.mllib.rdd中的类
-
:: DeveloperApi ::
Machine learning specific RDD functions.
- RDDFunctions(RDD<T>, ClassTag<T>) - 类 的构造器org.apache.spark.mllib.rdd.RDDFunctions
-
- rddId() - 类 中的方法org.apache.spark.CleanCheckpoint
-
- rddId() - 类 中的方法org.apache.spark.CleanRDD
-
- rddId() - 类 中的方法org.apache.spark.scheduler.SparkListenerUnpersistRDD
-
- rddId() - 类 中的方法org.apache.spark.storage.BlockManagerMessages.RemoveRdd
-
- rddId() - 类 中的方法org.apache.spark.storage.RDDBlockId
-
- rddIds() - 类 中的方法org.apache.spark.status.api.v1.StageData
-
- RDDInfo - org.apache.spark.storage中的类
-
- RDDInfo(int, String, int, StorageLevel, Seq<Object>, String, Option<org.apache.spark.rdd.RDDOperationScope>) - 类 的构造器org.apache.spark.storage.RDDInfo
-
- rddInfoFromJson(JsonAST.JValue) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- rddInfos() - 类 中的方法org.apache.spark.scheduler.StageInfo
-
- rddInfoToJson(RDDInfo) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- RDDPartitionInfo - org.apache.spark.status.api.v1中的类
-
- RDDPartitionSeq - org.apache.spark.status中的类
-
A custom sequence of partitions based on a mutable linked list.
- RDDPartitionSeq() - 类 的构造器org.apache.spark.status.RDDPartitionSeq
-
- rdds() - 类 中的方法org.apache.spark.rdd.CoGroupedRDD
-
- rdds() - 类 中的方法org.apache.spark.rdd.UnionRDD
-
- RDDStorageInfo - org.apache.spark.status.api.v1中的类
-
- rddStorageLevel(int) - 类 中的方法org.apache.spark.storage.StorageStatus
-
已过时。
Return the storage level, if any, used by the given RDD in this block manager.
- rddToAsyncRDDActions(RDD<T>, ClassTag<T>) - 类 中的静态方法org.apache.spark.rdd.RDD
-
- rddToDatasetHolder(RDD<T>, Encoder<T>) - 类 中的方法org.apache.spark.sql.SQLImplicits
-
- rddToOrderedRDDFunctions(RDD<Tuple2<K, V>>, Ordering<K>, ClassTag<K>, ClassTag<V>) - 类 中的静态方法org.apache.spark.rdd.RDD
-
- rddToPairRDDFunctions(RDD<Tuple2<K, V>>, ClassTag<K>, ClassTag<V>, Ordering<K>) - 类 中的静态方法org.apache.spark.rdd.RDD
-
- rddToSequenceFileRDDFunctions(RDD<Tuple2<K, V>>, ClassTag<K>, ClassTag<V>, <any>, <any>) - 类 中的静态方法org.apache.spark.rdd.RDD
-
- read() - 类 中的方法org.apache.spark.io.NioBufferedFileInputStream
-
- read(byte[], int, int) - 类 中的方法org.apache.spark.io.NioBufferedFileInputStream
-
- read() - 类 中的方法org.apache.spark.io.ReadAheadInputStream
-
- read(byte[], int, int) - 类 中的方法org.apache.spark.io.ReadAheadInputStream
-
- read() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- read() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- read() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- read() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- read() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- read() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- read() - 类 中的静态方法org.apache.spark.ml.classification.OneVsRest
-
- read() - 类 中的静态方法org.apache.spark.ml.classification.OneVsRestModel
-
- read() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- read() - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
- read() - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- read() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- read() - 类 中的静态方法org.apache.spark.ml.clustering.KMeansModel
-
- read() - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- read() - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- read() - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
-
- read() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- read() - 类 中的静态方法org.apache.spark.ml.feature.ColumnPruner
-
- read() - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizerModel
-
- read() - 类 中的静态方法org.apache.spark.ml.feature.IDFModel
-
- read() - 类 中的静态方法org.apache.spark.ml.feature.ImputerModel
-
- read() - 类 中的静态方法org.apache.spark.ml.feature.MaxAbsScalerModel
-
- read() - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSHModel
-
- read() - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScalerModel
-
- read() - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderModel
-
- read() - 类 中的静态方法org.apache.spark.ml.feature.PCAModel
-
- read() - 类 中的静态方法org.apache.spark.ml.feature.RFormulaModel
-
- read() - 类 中的静态方法org.apache.spark.ml.feature.StandardScalerModel
-
- read() - 类 中的静态方法org.apache.spark.ml.feature.StringIndexerModel
-
- read() - 类 中的静态方法org.apache.spark.ml.feature.VectorAttributeRewriter
-
- read() - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexerModel
-
- read() - 类 中的静态方法org.apache.spark.ml.feature.Word2VecModel
-
- read() - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowthModel
-
- read() - 类 中的静态方法org.apache.spark.ml.Pipeline
-
- read() - 类 中的静态方法org.apache.spark.ml.PipelineModel
-
- read() - 类 中的静态方法org.apache.spark.ml.recommendation.ALSModel
-
- read() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- read() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- read() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- read() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- read() - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegressionModel
-
- read() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- read() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- read() - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidator
-
- read() - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidatorModel
-
- read() - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplit
-
- read() - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplitModel
-
- read() - 接口 中的方法org.apache.spark.ml.util.DefaultParamsReadable
-
- read() - 接口 中的方法org.apache.spark.ml.util.MLReadable
-
Returns an MLReader instance for this class.
- read(Kryo, Input, Class<Iterable<?>>) - 类 中的方法org.apache.spark.serializer.JavaIterableWrapperSerializer
-
- read() - 类 中的方法org.apache.spark.sql.SparkSession
-
Returns a
DataFrameReader that can be used to read non-streaming data in as a
DataFrame.
- read() - 类 中的方法org.apache.spark.sql.SQLContext
-
- read() - 类 中的方法org.apache.spark.storage.BufferReleasingInputStream
-
- read(byte[]) - 类 中的方法org.apache.spark.storage.BufferReleasingInputStream
-
- read(byte[], int, int) - 类 中的方法org.apache.spark.storage.BufferReleasingInputStream
-
- read(String) - 类 中的静态方法org.apache.spark.streaming.CheckpointReader
-
Read checkpoint files present in the given checkpoint directory.
- read(String, SparkConf, Configuration, boolean) - 类 中的静态方法org.apache.spark.streaming.CheckpointReader
-
Read checkpoint files present in the given checkpoint directory.
- read(WriteAheadLogRecordHandle) - 类 中的方法org.apache.spark.streaming.util.WriteAheadLog
-
Read a written record based on the given record handle.
- read(Class<T>, Object) - 类 中的方法org.apache.spark.util.kvstore.InMemoryStore
-
- read(Class<T>, Object) - 接口 中的方法org.apache.spark.util.kvstore.KVStore
-
Read a specific instance of an object.
- read(Class<T>, Object) - 类 中的方法org.apache.spark.util.kvstore.LevelDB
-
- ReadableChannelFileRegion - org.apache.spark.storage中的类
-
- ReadableChannelFileRegion(ReadableByteChannel, long) - 类 的构造器org.apache.spark.storage.ReadableChannelFileRegion
-
- ReadAheadInputStream - org.apache.spark.io中的类
-
InputStream implementation which asynchronously reads ahead from the underlying input
stream when specified amount of data has been read from the current buffer.
- ReadAheadInputStream(InputStream, int, int) - 类 的构造器org.apache.spark.io.ReadAheadInputStream
-
Creates a ReadAheadInputStream with the specified buffer size and read-ahead
threshold
- readAll() - 类 中的方法org.apache.spark.streaming.util.WriteAheadLog
-
Read and return an iterator of all the records that have been written but not yet cleaned up.
- readArray(DataInputStream, JVMObjectTracker) - 类 中的静态方法org.apache.spark.api.r.SerDe
-
- readBoolean(DataInputStream) - 类 中的静态方法org.apache.spark.api.r.SerDe
-
- readBooleanArr(DataInputStream) - 类 中的静态方法org.apache.spark.api.r.SerDe
-
- readBytes(DataInputStream) - 类 中的静态方法org.apache.spark.api.r.SerDe
-
- readBytes() - 类 中的方法org.apache.spark.status.api.v1.ShuffleReadMetricDistributions
-
- readBytesArr(DataInputStream) - 类 中的静态方法org.apache.spark.api.r.SerDe
-
- readDate(DataInputStream) - 类 中的静态方法org.apache.spark.api.r.SerDe
-
- readDouble(DataInputStream) - 类 中的静态方法org.apache.spark.api.r.SerDe
-
- readDoubleArr(DataInputStream) - 类 中的静态方法org.apache.spark.api.r.SerDe
-
- readExternal(ObjectInput) - 类 中的方法org.apache.spark.serializer.JavaSerializer
-
- readExternal(ObjectInput) - 类 中的方法org.apache.spark.storage.BlockManagerId
-
- readExternal(ObjectInput) - 类 中的方法org.apache.spark.storage.BlockManagerMessages.UpdateBlockInfo
-
- readExternal(ObjectInput) - 类 中的方法org.apache.spark.storage.StorageLevel
-
- readFrom(ConfigReader) - 类 中的方法org.apache.spark.internal.config.ConfigEntryWithDefault
-
- readFrom(ConfigReader) - 类 中的方法org.apache.spark.internal.config.ConfigEntryWithDefaultFunction
-
- readFrom(ConfigReader) - 类 中的方法org.apache.spark.internal.config.ConfigEntryWithDefaultString
-
- readFrom(InputStream) - 类 中的静态方法org.apache.spark.util.sketch.BloomFilter
-
- readFrom(InputStream) - 类 中的静态方法org.apache.spark.util.sketch.CountMinSketch
-
- readFrom(byte[]) - 类 中的静态方法org.apache.spark.util.sketch.CountMinSketch
-
- readImages(String) - 类 中的静态方法org.apache.spark.ml.image.ImageSchema
-
Read the directory of images from the local or remote source
- readImages(String, SparkSession, boolean, int, boolean, double, long) - 类 中的静态方法org.apache.spark.ml.image.ImageSchema
-
Read the directory of images from the local or remote source
- readInt(DataInputStream) - 类 中的静态方法org.apache.spark.api.r.SerDe
-
- readIntArr(DataInputStream) - 类 中的静态方法org.apache.spark.api.r.SerDe
-
- readKey(ClassTag<T>) - 类 中的方法org.apache.spark.serializer.DeserializationStream
-
Reads the object representing the key of a key-value pair.
- readList(DataInputStream, JVMObjectTracker) - 类 中的静态方法org.apache.spark.api.r.SerDe
-
- readMap(DataInputStream, JVMObjectTracker) - 类 中的静态方法org.apache.spark.api.r.SerDe
-
- readObject(DataInputStream, JVMObjectTracker) - 类 中的静态方法org.apache.spark.api.r.SerDe
-
- readObject(ClassTag<T>) - 类 中的方法org.apache.spark.serializer.DeserializationStream
-
The most general-purpose method to read an object.
- readObjectType(DataInputStream) - 类 中的静态方法org.apache.spark.api.r.SerDe
-
- readRecords() - 类 中的方法org.apache.spark.status.api.v1.ShuffleReadMetricDistributions
-
- readSchema(Seq<String>, Option<Configuration>, boolean) - 类 中的静态方法org.apache.spark.sql.hive.orc.OrcFileOperator
-
- readSchema() - 接口 中的方法org.apache.spark.sql.sources.v2.reader.DataSourceReader
-
Returns the actual schema of this data source reader, which may be different from the physical
schema of the underlying storage, as column pruning or other optimizations may happen.
- readSqlObject(DataInputStream, char) - 类 中的静态方法org.apache.spark.sql.api.r.SQLUtils
-
- readStream() - 类 中的方法org.apache.spark.sql.SparkSession
-
Returns a DataStreamReader that can be used to read streaming data in as a DataFrame.
- readStream() - 类 中的方法org.apache.spark.sql.SQLContext
-
- readString(DataInputStream) - 类 中的静态方法org.apache.spark.api.r.SerDe
-
- readStringArr(DataInputStream) - 类 中的静态方法org.apache.spark.api.r.SerDe
-
- readStringBytes(DataInputStream, int) - 类 中的静态方法org.apache.spark.api.r.SerDe
-
- ReadSupport - org.apache.spark.sql.sources.v2中的接口
-
- ReadSupportWithSchema - org.apache.spark.sql.sources.v2中的接口
-
- readTime(DataInputStream) - 类 中的静态方法org.apache.spark.api.r.SerDe
-
- readTypedObject(DataInputStream, char, JVMObjectTracker) - 类 中的静态方法org.apache.spark.api.r.SerDe
-
- readValue(ClassTag<T>) - 类 中的方法org.apache.spark.serializer.DeserializationStream
-
Reads the object representing the value of a key-value pair.
- ready(Duration, CanAwait) - 类 中的方法org.apache.spark.ComplexFutureAction
-
- ready(Duration, CanAwait) - 接口 中的方法org.apache.spark.FutureAction
-
Blocks until this action completes.
- ready(Duration, CanAwait) - 类 中的方法org.apache.spark.SimpleFutureAction
-
- reason() - 类 中的方法org.apache.spark.ExecutorLostFailure
-
- reason() - 类 中的方法org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.KillTask
-
- reason() - 类 中的方法org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RemoveExecutor
-
- reason() - 类 中的方法org.apache.spark.scheduler.local.KillTask
-
- reason() - 类 中的方法org.apache.spark.scheduler.SparkListenerExecutorRemoved
-
- reason() - 类 中的方法org.apache.spark.scheduler.SparkListenerTaskEnd
-
- reason() - 类 中的方法org.apache.spark.TaskKilled
-
- reason() - 异常错误 中的方法org.apache.spark.TaskKilledException
-
- Recall - org.apache.spark.mllib.evaluation.binary中的类
-
Recall.
- Recall() - 类 的构造器org.apache.spark.mllib.evaluation.binary.Recall
-
- recall(double) - 类 中的方法org.apache.spark.mllib.evaluation.MulticlassMetrics
-
Returns recall for a given label (category)
- recall() - 类 中的方法org.apache.spark.mllib.evaluation.MulticlassMetrics
-
- recall() - 类 中的方法org.apache.spark.mllib.evaluation.MultilabelMetrics
-
Returns document-based recall averaged by the number of documents
- recall(double) - 类 中的方法org.apache.spark.mllib.evaluation.MultilabelMetrics
-
Returns recall for a given label (category)
- recallByLabel() - 接口 中的方法org.apache.spark.ml.classification.LogisticRegressionSummary
-
Returns recall for each label (category).
- recallByThreshold() - 接口 中的方法org.apache.spark.ml.classification.BinaryLogisticRegressionSummary
-
Returns a dataframe with two fields (threshold, recall) curve.
- recallByThreshold() - 类 中的方法org.apache.spark.mllib.evaluation.BinaryClassificationMetrics
-
Returns the (threshold, recall) curve.
- Receiver<T> - org.apache.spark.streaming.receiver中的类
-
:: DeveloperApi ::
Abstract class of a receiver that can be run on worker nodes to receive external data.
- Receiver(StorageLevel) - 类 的构造器org.apache.spark.streaming.receiver.Receiver
-
- RECEIVER_WAL_CLASS_CONF_KEY() - 类 中的静态方法org.apache.spark.streaming.util.WriteAheadLogUtils
-
- RECEIVER_WAL_CLOSE_AFTER_WRITE_CONF_KEY() - 类 中的静态方法org.apache.spark.streaming.util.WriteAheadLogUtils
-
- RECEIVER_WAL_ENABLE_CONF_KEY() - 类 中的静态方法org.apache.spark.streaming.util.WriteAheadLogUtils
-
- RECEIVER_WAL_MAX_FAILURES_CONF_KEY() - 类 中的静态方法org.apache.spark.streaming.util.WriteAheadLogUtils
-
- RECEIVER_WAL_ROLLING_INTERVAL_CONF_KEY() - 类 中的静态方法org.apache.spark.streaming.util.WriteAheadLogUtils
-
- ReceiverInfo - org.apache.spark.status.api.v1.streaming中的类
-
- ReceiverInfo - org.apache.spark.streaming.scheduler中的类
-
:: DeveloperApi ::
Class having information about a receiver
- ReceiverInfo(int, String, boolean, String, String, String, String, long) - 类 的构造器org.apache.spark.streaming.scheduler.ReceiverInfo
-
- receiverInfo() - 类 中的方法org.apache.spark.streaming.scheduler.StreamingListenerReceiverError
-
- receiverInfo() - 类 中的方法org.apache.spark.streaming.scheduler.StreamingListenerReceiverStarted
-
- receiverInfo() - 类 中的方法org.apache.spark.streaming.scheduler.StreamingListenerReceiverStopped
-
- receiverInputDStream() - 类 中的方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- receiverInputDStream() - 类 中的方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- ReceiverInputDStream<T> - org.apache.spark.streaming.dstream中的类
-
Abstract class for defining any
InputDStream
that has to start a receiver on worker nodes to receive external data.
- ReceiverInputDStream(StreamingContext, ClassTag<T>) - 类 的构造器org.apache.spark.streaming.dstream.ReceiverInputDStream
-
- ReceiverState - org.apache.spark.streaming.scheduler中的类
-
Enumeration to identify current state of a Receiver
- ReceiverState() - 类 的构造器org.apache.spark.streaming.scheduler.ReceiverState
-
- receiverStream(Receiver<T>) - 类 中的方法org.apache.spark.streaming.api.java.JavaStreamingContext
-
Create an input stream with any arbitrary user implemented receiver.
- receiverStream(Receiver<T>, ClassTag<T>) - 类 中的方法org.apache.spark.streaming.StreamingContext
-
Create an input stream with any arbitrary user implemented receiver.
- recentProgress() - 接口 中的方法org.apache.spark.sql.streaming.StreamingQuery
-
- recommendForAllItems(int) - 类 中的方法org.apache.spark.ml.recommendation.ALSModel
-
Returns top numUsers users recommended for each item, for all items.
- recommendForAllUsers(int) - 类 中的方法org.apache.spark.ml.recommendation.ALSModel
-
Returns top numItems items recommended for each user, for all users.
- recommendForItemSubset(Dataset<?>, int) - 类 中的方法org.apache.spark.ml.recommendation.ALSModel
-
Returns top numUsers users recommended for each item id in the input data set.
- recommendForUserSubset(Dataset<?>, int) - 类 中的方法org.apache.spark.ml.recommendation.ALSModel
-
Returns top numItems items recommended for each user id in the input data set.
- recommendProducts(int, int) - 类 中的方法org.apache.spark.mllib.recommendation.MatrixFactorizationModel
-
Recommends products to a user.
- recommendProductsForUsers(int) - 类 中的方法org.apache.spark.mllib.recommendation.MatrixFactorizationModel
-
Recommends top products for all users.
- recommendUsers(int, int) - 类 中的方法org.apache.spark.mllib.recommendation.MatrixFactorizationModel
-
Recommends users to a product.
- recommendUsersForProducts(int) - 类 中的方法org.apache.spark.mllib.recommendation.MatrixFactorizationModel
-
Recommends top users for all products.
- recordReader(InputStream, Configuration) - 类 中的方法org.apache.spark.sql.hive.execution.HiveScriptIOSchema
-
- recordReaderClass() - 类 中的方法org.apache.spark.sql.hive.execution.HiveScriptIOSchema
-
- RECORDS_BETWEEN_BYTES_READ_METRIC_UPDATES() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
Update the input bytes read metric each time this number of records has been read
- RECORDS_READ() - 类 中的方法org.apache.spark.InternalAccumulator.input$
-
- RECORDS_READ() - 类 中的方法org.apache.spark.InternalAccumulator.shuffleRead$
-
- RECORDS_WRITTEN() - 类 中的方法org.apache.spark.InternalAccumulator.output$
-
- RECORDS_WRITTEN() - 类 中的方法org.apache.spark.InternalAccumulator.shuffleWrite$
-
- recordsRead() - 类 中的方法org.apache.spark.status.api.v1.InputMetricDistributions
-
- recordsRead() - 类 中的方法org.apache.spark.status.api.v1.InputMetrics
-
- recordsRead() - 类 中的方法org.apache.spark.status.api.v1.ShuffleReadMetrics
-
- recordsWritten() - 类 中的方法org.apache.spark.status.api.v1.OutputMetricDistributions
-
- recordsWritten() - 类 中的方法org.apache.spark.status.api.v1.OutputMetrics
-
- recordsWritten() - 类 中的方法org.apache.spark.status.api.v1.ShuffleWriteMetrics
-
- recordWriter(OutputStream, Configuration) - 类 中的方法org.apache.spark.sql.hive.execution.HiveScriptIOSchema
-
- recordWriterClass() - 类 中的方法org.apache.spark.sql.hive.execution.HiveScriptIOSchema
-
- recoverPartitions(String) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
Recovers all the partitions in the directory of a table and update the catalog.
- RecursiveFlag - org.apache.spark.ml.image中的类
-
- RecursiveFlag() - 类 的构造器org.apache.spark.ml.image.RecursiveFlag
-
- recursiveList(File) - 类 中的静态方法org.apache.spark.util.Utils
-
Lists files recursively.
- redact(SparkConf, Seq<Tuple2<String, String>>) - 类 中的静态方法org.apache.spark.util.Utils
-
Redact the sensitive values in the given map.
- redact(Option<Regex>, Seq<Tuple2<String, String>>) - 类 中的静态方法org.apache.spark.util.Utils
-
Redact the sensitive values in the given map.
- redact(Option<Regex>, String) - 类 中的静态方法org.apache.spark.util.Utils
-
Redact the sensitive information in the given string.
- redact(Map<String, String>) - 类 中的静态方法org.apache.spark.util.Utils
-
Looks up the redaction regex from within the key value pairs and uses it to redact the rest
of the key value pairs.
- REDIRECT_CONNECTOR_NAME() - 类 中的静态方法org.apache.spark.ui.JettyUtils
-
- redirectError() - 类 中的方法org.apache.spark.launcher.SparkLauncher
-
Specifies that stderr in spark-submit should be redirected to stdout.
- redirectError(ProcessBuilder.Redirect) - 类 中的方法org.apache.spark.launcher.SparkLauncher
-
Redirects error output to the specified Redirect.
- redirectError(File) - 类 中的方法org.apache.spark.launcher.SparkLauncher
-
Redirects error output to the specified File.
- redirectOutput(ProcessBuilder.Redirect) - 类 中的方法org.apache.spark.launcher.SparkLauncher
-
Redirects standard output to the specified Redirect.
- redirectOutput(File) - 类 中的方法org.apache.spark.launcher.SparkLauncher
-
Redirects error output to the specified File.
- redirectToLog(String) - 类 中的方法org.apache.spark.launcher.SparkLauncher
-
Sets all output to be logged and redirected to a logger with the specified name.
- reduce(Function2<T, T, T>) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- reduce(Function2<T, T, T>) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- reduce(Function2<T, T, T>) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- reduce(Function2<T, T, T>) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Reduces the elements of this RDD using the specified commutative and associative binary
operator.
- reduce(Function2<T, T, T>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- reduce(Function2<T, T, T>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- reduce(Function2<T, T, T>) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- reduce(Function2<T, T, T>) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- reduce(Function2<T, T, T>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- reduce(Function2<T, T, T>) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- reduce(Function2<T, T, T>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- reduce(Function2<T, T, T>) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- reduce(Function2<T, T, T>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- reduce(Function2<T, T, T>) - 类 中的方法org.apache.spark.rdd.RDD
-
Reduces the elements of this RDD using the specified commutative and
associative binary operator.
- reduce(Function2<T, T, T>) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- reduce(Function2<T, T, T>) - 类 中的方法org.apache.spark.sql.Dataset
-
:: Experimental ::
(Scala-specific)
Reduces the elements of this Dataset using the specified binary function.
- reduce(ReduceFunction<T>) - 类 中的方法org.apache.spark.sql.Dataset
-
:: Experimental ::
(Java-specific)
Reduces the elements of this Dataset using the specified binary function.
- reduce(BUF, IN) - 类 中的方法org.apache.spark.sql.expressions.Aggregator
-
Combine two values to produce a new value.
- reduce(Function2<A1, A1, A1>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- reduce(Function2<T, T, T>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- reduce(Function2<T, T, T>) - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return a new DStream in which each RDD has a single element generated by reducing each RDD
of this DStream.
- reduce(Function2<T, T, T>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- reduce(Function2<T, T, T>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- reduce(Function2<T, T, T>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- reduce(Function2<T, T, T>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- reduce(Function2<T, T, T>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- reduce(Function2<T, T, T>) - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Return a new DStream in which each RDD has a single element generated by reducing each RDD
of this DStream.
- reduceByKey(Partitioner, Function2<V, V, V>) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Merge the values for each key using an associative and commutative reduce function.
- reduceByKey(Function2<V, V, V>, int) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Merge the values for each key using an associative and commutative reduce function.
- reduceByKey(Function2<V, V, V>) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Merge the values for each key using an associative and commutative reduce function.
- reduceByKey(Partitioner, Function2<V, V, V>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Merge the values for each key using an associative and commutative reduce function.
- reduceByKey(Function2<V, V, V>, int) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Merge the values for each key using an associative and commutative reduce function.
- reduceByKey(Function2<V, V, V>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Merge the values for each key using an associative and commutative reduce function.
- reduceByKey(Function2<V, V, V>) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying reduceByKey to each RDD.
- reduceByKey(Function2<V, V, V>, int) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying reduceByKey to each RDD.
- reduceByKey(Function2<V, V, V>, Partitioner) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying reduceByKey to each RDD.
- reduceByKey(Function2<V, V, V>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- reduceByKey(Function2<V, V, V>, int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- reduceByKey(Function2<V, V, V>, Partitioner) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- reduceByKey(Function2<V, V, V>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- reduceByKey(Function2<V, V, V>, int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- reduceByKey(Function2<V, V, V>, Partitioner) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- reduceByKey(Function2<V, V, V>) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying reduceByKey to each RDD.
- reduceByKey(Function2<V, V, V>, int) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying reduceByKey to each RDD.
- reduceByKey(Function2<V, V, V>, Partitioner) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying reduceByKey to each RDD.
- reduceByKeyAndWindow(Function2<V, V, V>, Duration) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Create a new DStream by applying reduceByKey over a sliding window on this DStream.
- reduceByKeyAndWindow(Function2<V, V, V>, Duration, Duration) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying reduceByKey over a sliding window.
- reduceByKeyAndWindow(Function2<V, V, V>, Duration, Duration, int) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying reduceByKey over a sliding window.
- reduceByKeyAndWindow(Function2<V, V, V>, Duration, Duration, Partitioner) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying reduceByKey over a sliding window.
- reduceByKeyAndWindow(Function2<V, V, V>, Function2<V, V, V>, Duration, Duration) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by reducing over a using incremental computation.
- reduceByKeyAndWindow(Function2<V, V, V>, Function2<V, V, V>, Duration, Duration, int, Function<Tuple2<K, V>, Boolean>) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying incremental reduceByKey over a sliding window.
- reduceByKeyAndWindow(Function2<V, V, V>, Function2<V, V, V>, Duration, Duration, Partitioner, Function<Tuple2<K, V>, Boolean>) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying incremental reduceByKey over a sliding window.
- reduceByKeyAndWindow(Function2<V, V, V>, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- reduceByKeyAndWindow(Function2<V, V, V>, Duration, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- reduceByKeyAndWindow(Function2<V, V, V>, Duration, Duration, int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- reduceByKeyAndWindow(Function2<V, V, V>, Duration, Duration, Partitioner) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- reduceByKeyAndWindow(Function2<V, V, V>, Function2<V, V, V>, Duration, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- reduceByKeyAndWindow(Function2<V, V, V>, Function2<V, V, V>, Duration, Duration, int, Function<Tuple2<K, V>, Boolean>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- reduceByKeyAndWindow(Function2<V, V, V>, Function2<V, V, V>, Duration, Duration, Partitioner, Function<Tuple2<K, V>, Boolean>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- reduceByKeyAndWindow(Function2<V, V, V>, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- reduceByKeyAndWindow(Function2<V, V, V>, Duration, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- reduceByKeyAndWindow(Function2<V, V, V>, Duration, Duration, int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- reduceByKeyAndWindow(Function2<V, V, V>, Duration, Duration, Partitioner) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- reduceByKeyAndWindow(Function2<V, V, V>, Function2<V, V, V>, Duration, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- reduceByKeyAndWindow(Function2<V, V, V>, Function2<V, V, V>, Duration, Duration, int, Function<Tuple2<K, V>, Boolean>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- reduceByKeyAndWindow(Function2<V, V, V>, Function2<V, V, V>, Duration, Duration, Partitioner, Function<Tuple2<K, V>, Boolean>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- reduceByKeyAndWindow(Function2<V, V, V>, Duration) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying reduceByKey over a sliding window on this DStream.
- reduceByKeyAndWindow(Function2<V, V, V>, Duration, Duration) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying reduceByKey over a sliding window.
- reduceByKeyAndWindow(Function2<V, V, V>, Duration, Duration, int) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying reduceByKey over a sliding window.
- reduceByKeyAndWindow(Function2<V, V, V>, Duration, Duration, Partitioner) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying reduceByKey over a sliding window.
- reduceByKeyAndWindow(Function2<V, V, V>, Function2<V, V, V>, Duration, Duration, int, Function1<Tuple2<K, V>, Object>) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying incremental reduceByKey over a sliding window.
- reduceByKeyAndWindow(Function2<V, V, V>, Function2<V, V, V>, Duration, Duration, Partitioner, Function1<Tuple2<K, V>, Object>) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying incremental reduceByKey over a sliding window.
- reduceByKeyLocally(Function2<V, V, V>) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Merge the values for each key using an associative and commutative reduce function, but return
the result immediately to the master as a Map.
- reduceByKeyLocally(Function2<V, V, V>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Merge the values for each key using an associative and commutative reduce function, but return
the results immediately to the master as a Map.
- reduceByWindow(Function2<T, T, T>, Duration, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- reduceByWindow(Function2<T, T, T>, Function2<T, T, T>, Duration, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- reduceByWindow(Function2<T, T, T>, Duration, Duration) - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return a new DStream in which each RDD has a single element generated by reducing all
elements in a sliding window over this DStream.
- reduceByWindow(Function2<T, T, T>, Function2<T, T, T>, Duration, Duration) - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return a new DStream in which each RDD has a single element generated by reducing all
elements in a sliding window over this DStream.
- reduceByWindow(Function2<T, T, T>, Duration, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- reduceByWindow(Function2<T, T, T>, Function2<T, T, T>, Duration, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- reduceByWindow(Function2<T, T, T>, Duration, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- reduceByWindow(Function2<T, T, T>, Function2<T, T, T>, Duration, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- reduceByWindow(Function2<T, T, T>, Duration, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- reduceByWindow(Function2<T, T, T>, Function2<T, T, T>, Duration, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- reduceByWindow(Function2<T, T, T>, Duration, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- reduceByWindow(Function2<T, T, T>, Function2<T, T, T>, Duration, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- reduceByWindow(Function2<T, T, T>, Duration, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- reduceByWindow(Function2<T, T, T>, Function2<T, T, T>, Duration, Duration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- reduceByWindow(Function2<T, T, T>, Duration, Duration) - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Return a new DStream in which each RDD has a single element generated by reducing all
elements in a sliding window over this DStream.
- reduceByWindow(Function2<T, T, T>, Function2<T, T, T>, Duration, Duration) - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Return a new DStream in which each RDD has a single element generated by reducing all
elements in a sliding window over this DStream.
- ReduceFunction<T> - org.apache.spark.api.java.function中的接口
-
Base interface for function used in Dataset's reduce.
- reduceGroups(Function2<V, V, V>) - 类 中的方法org.apache.spark.sql.KeyValueGroupedDataset
-
(Scala-specific)
Reduces the elements of each group of data using the specified binary function.
- reduceGroups(ReduceFunction<V>) - 类 中的方法org.apache.spark.sql.KeyValueGroupedDataset
-
(Java-specific)
Reduces the elements of each group of data using the specified binary function.
- reduceId() - 类 中的方法org.apache.spark.FetchFailed
-
- reduceId() - 类 中的方法org.apache.spark.storage.ShuffleBlockId
-
- reduceId() - 类 中的方法org.apache.spark.storage.ShuffleDataBlockId
-
- reduceId() - 类 中的方法org.apache.spark.storage.ShuffleIndexBlockId
-
- reduceLeft(Function2<B, A, B>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- reduceLeftOption(Function2<B, A, B>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- reduceOption(Function2<A1, A1, A1>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- reduceRight(Function2<A, B, B>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- reduceRightOption(Function2<A, B, B>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- references() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- references() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- references() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- references() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- references() - 类 中的方法org.apache.spark.sql.sources.And
-
- references() - 类 中的方法org.apache.spark.sql.sources.EqualNullSafe
-
- references() - 类 中的方法org.apache.spark.sql.sources.EqualTo
-
- references() - 类 中的方法org.apache.spark.sql.sources.Filter
-
List of columns that are referenced by this filter.
- references() - 类 中的方法org.apache.spark.sql.sources.GreaterThan
-
- references() - 类 中的方法org.apache.spark.sql.sources.GreaterThanOrEqual
-
- references() - 类 中的方法org.apache.spark.sql.sources.In
-
- references() - 类 中的方法org.apache.spark.sql.sources.IsNotNull
-
- references() - 类 中的方法org.apache.spark.sql.sources.IsNull
-
- references() - 类 中的方法org.apache.spark.sql.sources.LessThan
-
- references() - 类 中的方法org.apache.spark.sql.sources.LessThanOrEqual
-
- references() - 类 中的方法org.apache.spark.sql.sources.Not
-
- references() - 类 中的方法org.apache.spark.sql.sources.Or
-
- references() - 类 中的方法org.apache.spark.sql.sources.StringContains
-
- references() - 类 中的方法org.apache.spark.sql.sources.StringEndsWith
-
- references() - 类 中的方法org.apache.spark.sql.sources.StringStartsWith
-
- refresh() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- refresh() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- refresh() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- refreshByPath(String) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
Invalidates and refreshes all the cached data (and the associated metadata) for any Dataset
that contains the given data source path.
- refreshTable(String) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
Invalidates and refreshes all the cached data and metadata of the given table.
- refreshTable(String) - 类 中的方法org.apache.spark.sql.hive.HiveContext
-
已过时。
Invalidate and refresh all the cached the metadata of the given table.
- regex(Regex) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaParser
-
- regexFromString(String, String) - 类 中的静态方法org.apache.spark.internal.config.ConfigHelpers
-
- regexp_extract(Column, String, int) - 类 中的静态方法org.apache.spark.sql.functions
-
Extract a specific group matched by a Java regex, from the specified string column.
- regexp_replace(Column, String, String) - 类 中的静态方法org.apache.spark.sql.functions
-
Replace all substrings of the specified string value that match regexp with rep.
- regexp_replace(Column, Column, Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Replace all substrings of the specified string value that match regexp with rep.
- RegexTokenizer - org.apache.spark.ml.feature中的类
-
A regex based tokenizer that extracts tokens either by using the provided regex pattern to split
the text (default) or repeatedly matching the regex (if gaps is false).
- RegexTokenizer(String) - 类 的构造器org.apache.spark.ml.feature.RegexTokenizer
-
- RegexTokenizer() - 类 的构造器org.apache.spark.ml.feature.RegexTokenizer
-
- register(AccumulatorV2<?, ?>) - 类 中的方法org.apache.spark.SparkContext
-
Register the given accumulator.
- register(AccumulatorV2<?, ?>, String) - 类 中的方法org.apache.spark.SparkContext
-
Register the given accumulator with given name.
- register(String, String) - 类 中的静态方法org.apache.spark.sql.types.UDTRegistration
-
Registers an UserDefinedType to an user class.
- register(String, UserDefinedAggregateFunction) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Registers a user-defined aggregate function (UDAF).
- register(String, UserDefinedFunction) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Registers a user-defined function (UDF), for a UDF that's already defined using the Dataset
API (i.e. of type UserDefinedFunction).
- register(String, Function0<RT>, TypeTags.TypeTag<RT>) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Registers a deterministic Scala closure of 0 arguments as user-defined function (UDF).
- register(String, Function1<A1, RT>, TypeTags.TypeTag<RT>, TypeTags.TypeTag<A1>) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Registers a deterministic Scala closure of 1 arguments as user-defined function (UDF).
- register(String, Function2<A1, A2, RT>, TypeTags.TypeTag<RT>, TypeTags.TypeTag<A1>, TypeTags.TypeTag<A2>) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Registers a deterministic Scala closure of 2 arguments as user-defined function (UDF).
- register(String, Function3<A1, A2, A3, RT>, TypeTags.TypeTag<RT>, TypeTags.TypeTag<A1>, TypeTags.TypeTag<A2>, TypeTags.TypeTag<A3>) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Registers a deterministic Scala closure of 3 arguments as user-defined function (UDF).
- register(String, Function4<A1, A2, A3, A4, RT>, TypeTags.TypeTag<RT>, TypeTags.TypeTag<A1>, TypeTags.TypeTag<A2>, TypeTags.TypeTag<A3>, TypeTags.TypeTag<A4>) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Registers a deterministic Scala closure of 4 arguments as user-defined function (UDF).
- register(String, Function5<A1, A2, A3, A4, A5, RT>, TypeTags.TypeTag<RT>, TypeTags.TypeTag<A1>, TypeTags.TypeTag<A2>, TypeTags.TypeTag<A3>, TypeTags.TypeTag<A4>, TypeTags.TypeTag<A5>) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Registers a deterministic Scala closure of 5 arguments as user-defined function (UDF).
- register(String, Function6<A1, A2, A3, A4, A5, A6, RT>, TypeTags.TypeTag<RT>, TypeTags.TypeTag<A1>, TypeTags.TypeTag<A2>, TypeTags.TypeTag<A3>, TypeTags.TypeTag<A4>, TypeTags.TypeTag<A5>, TypeTags.TypeTag<A6>) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Registers a deterministic Scala closure of 6 arguments as user-defined function (UDF).
- register(String, Function7<A1, A2, A3, A4, A5, A6, A7, RT>, TypeTags.TypeTag<RT>, TypeTags.TypeTag<A1>, TypeTags.TypeTag<A2>, TypeTags.TypeTag<A3>, TypeTags.TypeTag<A4>, TypeTags.TypeTag<A5>, TypeTags.TypeTag<A6>, TypeTags.TypeTag<A7>) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Registers a deterministic Scala closure of 7 arguments as user-defined function (UDF).
- register(String, Function8<A1, A2, A3, A4, A5, A6, A7, A8, RT>, TypeTags.TypeTag<RT>, TypeTags.TypeTag<A1>, TypeTags.TypeTag<A2>, TypeTags.TypeTag<A3>, TypeTags.TypeTag<A4>, TypeTags.TypeTag<A5>, TypeTags.TypeTag<A6>, TypeTags.TypeTag<A7>, TypeTags.TypeTag<A8>) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Registers a deterministic Scala closure of 8 arguments as user-defined function (UDF).
- register(String, Function9<A1, A2, A3, A4, A5, A6, A7, A8, A9, RT>, TypeTags.TypeTag<RT>, TypeTags.TypeTag<A1>, TypeTags.TypeTag<A2>, TypeTags.TypeTag<A3>, TypeTags.TypeTag<A4>, TypeTags.TypeTag<A5>, TypeTags.TypeTag<A6>, TypeTags.TypeTag<A7>, TypeTags.TypeTag<A8>, TypeTags.TypeTag<A9>) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Registers a deterministic Scala closure of 9 arguments as user-defined function (UDF).
- register(String, Function10<A1, A2, A3, A4, A5, A6, A7, A8, A9, A10, RT>, TypeTags.TypeTag<RT>, TypeTags.TypeTag<A1>, TypeTags.TypeTag<A2>, TypeTags.TypeTag<A3>, TypeTags.TypeTag<A4>, TypeTags.TypeTag<A5>, TypeTags.TypeTag<A6>, TypeTags.TypeTag<A7>, TypeTags.TypeTag<A8>, TypeTags.TypeTag<A9>, TypeTags.TypeTag<A10>) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Registers a deterministic Scala closure of 10 arguments as user-defined function (UDF).
- register(String, Function11<A1, A2, A3, A4, A5, A6, A7, A8, A9, A10, A11, RT>, TypeTags.TypeTag<RT>, TypeTags.TypeTag<A1>, TypeTags.TypeTag<A2>, TypeTags.TypeTag<A3>, TypeTags.TypeTag<A4>, TypeTags.TypeTag<A5>, TypeTags.TypeTag<A6>, TypeTags.TypeTag<A7>, TypeTags.TypeTag<A8>, TypeTags.TypeTag<A9>, TypeTags.TypeTag<A10>, TypeTags.TypeTag<A11>) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Registers a deterministic Scala closure of 11 arguments as user-defined function (UDF).
- register(String, Function12<A1, A2, A3, A4, A5, A6, A7, A8, A9, A10, A11, A12, RT>, TypeTags.TypeTag<RT>, TypeTags.TypeTag<A1>, TypeTags.TypeTag<A2>, TypeTags.TypeTag<A3>, TypeTags.TypeTag<A4>, TypeTags.TypeTag<A5>, TypeTags.TypeTag<A6>, TypeTags.TypeTag<A7>, TypeTags.TypeTag<A8>, TypeTags.TypeTag<A9>, TypeTags.TypeTag<A10>, TypeTags.TypeTag<A11>, TypeTags.TypeTag<A12>) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Registers a deterministic Scala closure of 12 arguments as user-defined function (UDF).
- register(String, Function13<A1, A2, A3, A4, A5, A6, A7, A8, A9, A10, A11, A12, A13, RT>, TypeTags.TypeTag<RT>, TypeTags.TypeTag<A1>, TypeTags.TypeTag<A2>, TypeTags.TypeTag<A3>, TypeTags.TypeTag<A4>, TypeTags.TypeTag<A5>, TypeTags.TypeTag<A6>, TypeTags.TypeTag<A7>, TypeTags.TypeTag<A8>, TypeTags.TypeTag<A9>, TypeTags.TypeTag<A10>, TypeTags.TypeTag<A11>, TypeTags.TypeTag<A12>, TypeTags.TypeTag<A13>) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Registers a deterministic Scala closure of 13 arguments as user-defined function (UDF).
- register(String, Function14<A1, A2, A3, A4, A5, A6, A7, A8, A9, A10, A11, A12, A13, A14, RT>, TypeTags.TypeTag<RT>, TypeTags.TypeTag<A1>, TypeTags.TypeTag<A2>, TypeTags.TypeTag<A3>, TypeTags.TypeTag<A4>, TypeTags.TypeTag<A5>, TypeTags.TypeTag<A6>, TypeTags.TypeTag<A7>, TypeTags.TypeTag<A8>, TypeTags.TypeTag<A9>, TypeTags.TypeTag<A10>, TypeTags.TypeTag<A11>, TypeTags.TypeTag<A12>, TypeTags.TypeTag<A13>, TypeTags.TypeTag<A14>) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Registers a deterministic Scala closure of 14 arguments as user-defined function (UDF).
- register(String, Function15<A1, A2, A3, A4, A5, A6, A7, A8, A9, A10, A11, A12, A13, A14, A15, RT>, TypeTags.TypeTag<RT>, TypeTags.TypeTag<A1>, TypeTags.TypeTag<A2>, TypeTags.TypeTag<A3>, TypeTags.TypeTag<A4>, TypeTags.TypeTag<A5>, TypeTags.TypeTag<A6>, TypeTags.TypeTag<A7>, TypeTags.TypeTag<A8>, TypeTags.TypeTag<A9>, TypeTags.TypeTag<A10>, TypeTags.TypeTag<A11>, TypeTags.TypeTag<A12>, TypeTags.TypeTag<A13>, TypeTags.TypeTag<A14>, TypeTags.TypeTag<A15>) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Registers a deterministic Scala closure of 15 arguments as user-defined function (UDF).
- register(String, Function16<A1, A2, A3, A4, A5, A6, A7, A8, A9, A10, A11, A12, A13, A14, A15, A16, RT>, TypeTags.TypeTag<RT>, TypeTags.TypeTag<A1>, TypeTags.TypeTag<A2>, TypeTags.TypeTag<A3>, TypeTags.TypeTag<A4>, TypeTags.TypeTag<A5>, TypeTags.TypeTag<A6>, TypeTags.TypeTag<A7>, TypeTags.TypeTag<A8>, TypeTags.TypeTag<A9>, TypeTags.TypeTag<A10>, TypeTags.TypeTag<A11>, TypeTags.TypeTag<A12>, TypeTags.TypeTag<A13>, TypeTags.TypeTag<A14>, TypeTags.TypeTag<A15>, TypeTags.TypeTag<A16>) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Registers a deterministic Scala closure of 16 arguments as user-defined function (UDF).
- register(String, Function17<A1, A2, A3, A4, A5, A6, A7, A8, A9, A10, A11, A12, A13, A14, A15, A16, A17, RT>, TypeTags.TypeTag<RT>, TypeTags.TypeTag<A1>, TypeTags.TypeTag<A2>, TypeTags.TypeTag<A3>, TypeTags.TypeTag<A4>, TypeTags.TypeTag<A5>, TypeTags.TypeTag<A6>, TypeTags.TypeTag<A7>, TypeTags.TypeTag<A8>, TypeTags.TypeTag<A9>, TypeTags.TypeTag<A10>, TypeTags.TypeTag<A11>, TypeTags.TypeTag<A12>, TypeTags.TypeTag<A13>, TypeTags.TypeTag<A14>, TypeTags.TypeTag<A15>, TypeTags.TypeTag<A16>, TypeTags.TypeTag<A17>) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Registers a deterministic Scala closure of 17 arguments as user-defined function (UDF).
- register(String, Function18<A1, A2, A3, A4, A5, A6, A7, A8, A9, A10, A11, A12, A13, A14, A15, A16, A17, A18, RT>, TypeTags.TypeTag<RT>, TypeTags.TypeTag<A1>, TypeTags.TypeTag<A2>, TypeTags.TypeTag<A3>, TypeTags.TypeTag<A4>, TypeTags.TypeTag<A5>, TypeTags.TypeTag<A6>, TypeTags.TypeTag<A7>, TypeTags.TypeTag<A8>, TypeTags.TypeTag<A9>, TypeTags.TypeTag<A10>, TypeTags.TypeTag<A11>, TypeTags.TypeTag<A12>, TypeTags.TypeTag<A13>, TypeTags.TypeTag<A14>, TypeTags.TypeTag<A15>, TypeTags.TypeTag<A16>, TypeTags.TypeTag<A17>, TypeTags.TypeTag<A18>) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Registers a deterministic Scala closure of 18 arguments as user-defined function (UDF).
- register(String, Function19<A1, A2, A3, A4, A5, A6, A7, A8, A9, A10, A11, A12, A13, A14, A15, A16, A17, A18, A19, RT>, TypeTags.TypeTag<RT>, TypeTags.TypeTag<A1>, TypeTags.TypeTag<A2>, TypeTags.TypeTag<A3>, TypeTags.TypeTag<A4>, TypeTags.TypeTag<A5>, TypeTags.TypeTag<A6>, TypeTags.TypeTag<A7>, TypeTags.TypeTag<A8>, TypeTags.TypeTag<A9>, TypeTags.TypeTag<A10>, TypeTags.TypeTag<A11>, TypeTags.TypeTag<A12>, TypeTags.TypeTag<A13>, TypeTags.TypeTag<A14>, TypeTags.TypeTag<A15>, TypeTags.TypeTag<A16>, TypeTags.TypeTag<A17>, TypeTags.TypeTag<A18>, TypeTags.TypeTag<A19>) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Registers a deterministic Scala closure of 19 arguments as user-defined function (UDF).
- register(String, Function20<A1, A2, A3, A4, A5, A6, A7, A8, A9, A10, A11, A12, A13, A14, A15, A16, A17, A18, A19, A20, RT>, TypeTags.TypeTag<RT>, TypeTags.TypeTag<A1>, TypeTags.TypeTag<A2>, TypeTags.TypeTag<A3>, TypeTags.TypeTag<A4>, TypeTags.TypeTag<A5>, TypeTags.TypeTag<A6>, TypeTags.TypeTag<A7>, TypeTags.TypeTag<A8>, TypeTags.TypeTag<A9>, TypeTags.TypeTag<A10>, TypeTags.TypeTag<A11>, TypeTags.TypeTag<A12>, TypeTags.TypeTag<A13>, TypeTags.TypeTag<A14>, TypeTags.TypeTag<A15>, TypeTags.TypeTag<A16>, TypeTags.TypeTag<A17>, TypeTags.TypeTag<A18>, TypeTags.TypeTag<A19>, TypeTags.TypeTag<A20>) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Registers a deterministic Scala closure of 20 arguments as user-defined function (UDF).
- register(String, Function21<A1, A2, A3, A4, A5, A6, A7, A8, A9, A10, A11, A12, A13, A14, A15, A16, A17, A18, A19, A20, A21, RT>, TypeTags.TypeTag<RT>, TypeTags.TypeTag<A1>, TypeTags.TypeTag<A2>, TypeTags.TypeTag<A3>, TypeTags.TypeTag<A4>, TypeTags.TypeTag<A5>, TypeTags.TypeTag<A6>, TypeTags.TypeTag<A7>, TypeTags.TypeTag<A8>, TypeTags.TypeTag<A9>, TypeTags.TypeTag<A10>, TypeTags.TypeTag<A11>, TypeTags.TypeTag<A12>, TypeTags.TypeTag<A13>, TypeTags.TypeTag<A14>, TypeTags.TypeTag<A15>, TypeTags.TypeTag<A16>, TypeTags.TypeTag<A17>, TypeTags.TypeTag<A18>, TypeTags.TypeTag<A19>, TypeTags.TypeTag<A20>, TypeTags.TypeTag<A21>) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Registers a deterministic Scala closure of 21 arguments as user-defined function (UDF).
- register(String, Function22<A1, A2, A3, A4, A5, A6, A7, A8, A9, A10, A11, A12, A13, A14, A15, A16, A17, A18, A19, A20, A21, A22, RT>, TypeTags.TypeTag<RT>, TypeTags.TypeTag<A1>, TypeTags.TypeTag<A2>, TypeTags.TypeTag<A3>, TypeTags.TypeTag<A4>, TypeTags.TypeTag<A5>, TypeTags.TypeTag<A6>, TypeTags.TypeTag<A7>, TypeTags.TypeTag<A8>, TypeTags.TypeTag<A9>, TypeTags.TypeTag<A10>, TypeTags.TypeTag<A11>, TypeTags.TypeTag<A12>, TypeTags.TypeTag<A13>, TypeTags.TypeTag<A14>, TypeTags.TypeTag<A15>, TypeTags.TypeTag<A16>, TypeTags.TypeTag<A17>, TypeTags.TypeTag<A18>, TypeTags.TypeTag<A19>, TypeTags.TypeTag<A20>, TypeTags.TypeTag<A21>, TypeTags.TypeTag<A22>) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Registers a deterministic Scala closure of 22 arguments as user-defined function (UDF).
- register(String, UDF0<?>, DataType) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Register a deterministic Java UDF0 instance as user-defined function (UDF).
- register(String, UDF1<?, ?>, DataType) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Register a deterministic Java UDF1 instance as user-defined function (UDF).
- register(String, UDF2<?, ?, ?>, DataType) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Register a deterministic Java UDF2 instance as user-defined function (UDF).
- register(String, UDF3<?, ?, ?, ?>, DataType) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Register a deterministic Java UDF3 instance as user-defined function (UDF).
- register(String, UDF4<?, ?, ?, ?, ?>, DataType) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Register a deterministic Java UDF4 instance as user-defined function (UDF).
- register(String, UDF5<?, ?, ?, ?, ?, ?>, DataType) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Register a deterministic Java UDF5 instance as user-defined function (UDF).
- register(String, UDF6<?, ?, ?, ?, ?, ?, ?>, DataType) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Register a deterministic Java UDF6 instance as user-defined function (UDF).
- register(String, UDF7<?, ?, ?, ?, ?, ?, ?, ?>, DataType) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Register a deterministic Java UDF7 instance as user-defined function (UDF).
- register(String, UDF8<?, ?, ?, ?, ?, ?, ?, ?, ?>, DataType) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Register a deterministic Java UDF8 instance as user-defined function (UDF).
- register(String, UDF9<?, ?, ?, ?, ?, ?, ?, ?, ?, ?>, DataType) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Register a deterministic Java UDF9 instance as user-defined function (UDF).
- register(String, UDF10<?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?>, DataType) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Register a deterministic Java UDF10 instance as user-defined function (UDF).
- register(String, UDF11<?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?>, DataType) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Register a deterministic Java UDF11 instance as user-defined function (UDF).
- register(String, UDF12<?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?>, DataType) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Register a deterministic Java UDF12 instance as user-defined function (UDF).
- register(String, UDF13<?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?>, DataType) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Register a deterministic Java UDF13 instance as user-defined function (UDF).
- register(String, UDF14<?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?>, DataType) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Register a deterministic Java UDF14 instance as user-defined function (UDF).
- register(String, UDF15<?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?>, DataType) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Register a deterministic Java UDF15 instance as user-defined function (UDF).
- register(String, UDF16<?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?>, DataType) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Register a deterministic Java UDF16 instance as user-defined function (UDF).
- register(String, UDF17<?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?>, DataType) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Register a deterministic Java UDF17 instance as user-defined function (UDF).
- register(String, UDF18<?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?>, DataType) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Register a deterministic Java UDF18 instance as user-defined function (UDF).
- register(String, UDF19<?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?>, DataType) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Register a deterministic Java UDF19 instance as user-defined function (UDF).
- register(String, UDF20<?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?>, DataType) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Register a deterministic Java UDF20 instance as user-defined function (UDF).
- register(String, UDF21<?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?>, DataType) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Register a deterministic Java UDF21 instance as user-defined function (UDF).
- register(String, UDF22<?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?>, DataType) - 类 中的方法org.apache.spark.sql.UDFRegistration
-
Register a deterministic Java UDF22 instance as user-defined function (UDF).
- register(QueryExecutionListener) - 类 中的方法org.apache.spark.sql.util.ExecutionListenerManager
-
- register(AccumulatorV2<?, ?>) - 类 中的静态方法org.apache.spark.util.AccumulatorContext
-
Registers an
AccumulatorV2 created on the driver such that it can be used on the executors.
- register(String, Function0<Object>) - 类 中的静态方法org.apache.spark.util.SignalUtils
-
Adds an action to be run when a given signal is received by this process.
- registerAvroSchemas(Seq<Schema>) - 类 中的方法org.apache.spark.SparkConf
-
Use Kryo serialization and register the given set of Avro schemas so that the generic
record serializer can decrease network IO
- RegisterBlockManager(BlockManagerId, long, long, org.apache.spark.rpc.RpcEndpointRef) - 类 的构造器org.apache.spark.storage.BlockManagerMessages.RegisterBlockManager
-
- RegisterBlockManager$() - 类 的构造器org.apache.spark.storage.BlockManagerMessages.RegisterBlockManager$
-
- registerClasses(Kryo) - 接口 中的方法org.apache.spark.serializer.KryoRegistrator
-
- RegisterClusterManager(org.apache.spark.rpc.RpcEndpointRef) - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RegisterClusterManager
-
- RegisterClusterManager$() - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RegisterClusterManager$
-
- registerDialect(JdbcDialect) - 类 中的静态方法org.apache.spark.sql.jdbc.JdbcDialects
-
Register a dialect for use on all new matching jdbc org.apache.spark.sql.DataFrame.
- RegisteredExecutor$() - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RegisteredExecutor$
-
- RegisterExecutor(String, org.apache.spark.rpc.RpcEndpointRef, String, int, Map<String, String>) - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RegisterExecutor
-
- RegisterExecutor$() - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RegisterExecutor$
-
- RegisterExecutorFailed(String) - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RegisterExecutorFailed
-
- RegisterExecutorFailed$() - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RegisterExecutorFailed$
-
- registerKryoClasses(SparkConf) - 类 中的静态方法org.apache.spark.graphx.GraphXUtils
-
Registers classes that GraphX uses with Kryo.
- registerKryoClasses(SparkContext) - 类 中的静态方法org.apache.spark.ml.evaluation.SquaredEuclideanSilhouette
-
- registerKryoClasses(Class<?>[]) - 类 中的方法org.apache.spark.SparkConf
-
Use Kryo serialization and register the given set of classes with Kryo.
- registerLogger(Logger) - 类 中的静态方法org.apache.spark.util.SignalUtils
-
Register a signal handler to log signals on UNIX-like systems.
- registerShutdownDeleteDir(File) - 类 中的静态方法org.apache.spark.util.ShutdownHookManager
-
- registerStream(DStream<BinarySample>) - 类 中的方法org.apache.spark.mllib.stat.test.StreamingTest
-
Register a DStream of values for significance testing.
- registerStream(JavaDStream<BinarySample>) - 类 中的方法org.apache.spark.mllib.stat.test.StreamingTest
-
Register a JavaDStream of values for significance testing.
- registerTempTable(String) - 类 中的方法org.apache.spark.sql.Dataset
-
- regParam() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- regParam() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- regParam() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- regParam() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- regParam() - 接口 中的方法org.apache.spark.ml.param.shared.HasRegParam
-
Param for regularization parameter (>= 0).
- regParam() - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- regParam() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- regParam() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- regParam() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- regParam() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- Regression() - 类 中的静态方法org.apache.spark.mllib.tree.configuration.Algo
-
- RegressionEvaluator - org.apache.spark.ml.evaluation中的类
-
:: Experimental ::
Evaluator for regression, which expects two input columns: prediction and label.
- RegressionEvaluator(String) - 类 的构造器org.apache.spark.ml.evaluation.RegressionEvaluator
-
- RegressionEvaluator() - 类 的构造器org.apache.spark.ml.evaluation.RegressionEvaluator
-
- RegressionMetrics - org.apache.spark.mllib.evaluation中的类
-
Evaluator for regression.
- RegressionMetrics(RDD<Tuple2<Object, Object>>, boolean) - 类 的构造器org.apache.spark.mllib.evaluation.RegressionMetrics
-
- RegressionMetrics(RDD<Tuple2<Object, Object>>) - 类 的构造器org.apache.spark.mllib.evaluation.RegressionMetrics
-
- RegressionModel<FeaturesType,M extends RegressionModel<FeaturesType,M>> - org.apache.spark.ml.regression中的类
-
:: DeveloperApi ::
Model produced by a Regressor.
- RegressionModel() - 类 的构造器org.apache.spark.ml.regression.RegressionModel
-
- RegressionModel - org.apache.spark.mllib.regression中的接口
-
- reindex() - 类 中的方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- reindex() - 类 中的方法org.apache.spark.graphx.VertexRDD
-
Construct a new VertexRDD that is indexed by only the visible vertices.
- RelationalGroupedDataset - org.apache.spark.sql中的类
-
A set of methods for aggregations on a
DataFrame, created by
groupBy,
cube or
rollup (and also
pivot).
- RelationalGroupedDataset.CubeType$ - org.apache.spark.sql中的类
-
To indicate it's the CUBE
- RelationalGroupedDataset.GroupByType$ - org.apache.spark.sql中的类
-
To indicate it's the GroupBy
- RelationalGroupedDataset.PivotType$ - org.apache.spark.sql中的类
-
- RelationalGroupedDataset.RollupType$ - org.apache.spark.sql中的类
-
To indicate it's the ROLLUP
- RelationConversions - org.apache.spark.sql.hive中的类
-
Relation conversion from metastore relations to data source relations for better performance
- When writing to non-partitioned Hive-serde Parquet/Orc tables
- When scanning Hive-serde Parquet/ORC tables
This rule must be run before all other DDL post-hoc resolution rules, i.e.
- RelationConversions(SQLConf, HiveSessionCatalog) - 类 的构造器org.apache.spark.sql.hive.RelationConversions
-
- RelationProvider - org.apache.spark.sql.sources中的接口
-
Implemented by objects that produce relations for a specific kind of data source.
- relativeDirection(long) - 类 中的方法org.apache.spark.graphx.Edge
-
Return the relative direction of the edge to the corresponding
vertex.
- relativeError() - 类 中的静态方法org.apache.spark.ml.feature.QuantileDiscretizer
-
- relativeError() - 类 中的方法org.apache.spark.util.sketch.CountMinSketch
-
- rem(Decimal, Decimal) - 类 中的方法org.apache.spark.sql.types.Decimal.DecimalAsIfIntegral$
-
- remainder(Decimal) - 类 中的方法org.apache.spark.sql.types.Decimal
-
- remember(Duration) - 类 中的方法org.apache.spark.streaming.api.java.JavaStreamingContext
-
Sets each DStreams in this context to remember RDDs it generated in the last given duration.
- remember(Duration) - 类 中的方法org.apache.spark.streaming.StreamingContext
-
Set each DStream in this context to remember RDDs it generated in the last given duration.
- REMOTE_BLOCKS_FETCHED() - 类 中的方法org.apache.spark.InternalAccumulator.shuffleRead$
-
- REMOTE_BYTES_READ() - 类 中的方法org.apache.spark.InternalAccumulator.shuffleRead$
-
- REMOTE_BYTES_READ_TO_DISK() - 类 中的方法org.apache.spark.InternalAccumulator.shuffleRead$
-
- remoteBlocksFetched() - 类 中的方法org.apache.spark.status.api.v1.ShuffleReadMetricDistributions
-
- remoteBlocksFetched() - 类 中的方法org.apache.spark.status.api.v1.ShuffleReadMetrics
-
- remoteBytesRead() - 类 中的方法org.apache.spark.status.api.v1.ShuffleReadMetricDistributions
-
- remoteBytesRead() - 类 中的方法org.apache.spark.status.api.v1.ShuffleReadMetrics
-
- remoteBytesReadToDisk() - 类 中的方法org.apache.spark.status.api.v1.ShuffleReadMetricDistributions
-
- remoteBytesReadToDisk() - 类 中的方法org.apache.spark.status.api.v1.ShuffleReadMetrics
-
- remove(Param<T>) - 类 中的方法org.apache.spark.ml.param.ParamMap
-
Removes a key from this map and returns its value associated previously as an option.
- remove(String) - 类 中的方法org.apache.spark.SparkConf
-
Remove a parameter from the configuration
- remove() - 接口 中的方法org.apache.spark.sql.streaming.GroupState
-
Remove this state.
- remove(String) - 类 中的方法org.apache.spark.sql.types.MetadataBuilder
-
- remove() - 类 中的方法org.apache.spark.streaming.State
-
Remove the state if it exists.
- remove(long) - 类 中的静态方法org.apache.spark.util.AccumulatorContext
-
- RemoveBlock(BlockId) - 类 的构造器org.apache.spark.storage.BlockManagerMessages.RemoveBlock
-
- RemoveBlock$() - 类 的构造器org.apache.spark.storage.BlockManagerMessages.RemoveBlock$
-
- RemoveBroadcast(long, boolean) - 类 的构造器org.apache.spark.storage.BlockManagerMessages.RemoveBroadcast
-
- RemoveBroadcast$() - 类 的构造器org.apache.spark.storage.BlockManagerMessages.RemoveBroadcast$
-
- removeDistribution(LiveExecutor) - 类 中的方法org.apache.spark.status.LiveRDD
-
- RemoveExecutor(String, org.apache.spark.scheduler.ExecutorLossReason) - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RemoveExecutor
-
- RemoveExecutor(String) - 类 的构造器org.apache.spark.storage.BlockManagerMessages.RemoveExecutor
-
- RemoveExecutor$() - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RemoveExecutor$
-
- RemoveExecutor$() - 类 的构造器org.apache.spark.storage.BlockManagerMessages.RemoveExecutor$
-
- removeFromDriver() - 类 中的方法org.apache.spark.storage.BlockManagerMessages.RemoveBroadcast
-
- removeListener(L) - 类 中的静态方法org.apache.spark.scheduler.AsyncEventQueue
-
- removeListener(StreamingQueryListener) - 类 中的方法org.apache.spark.sql.streaming.StreamingQueryManager
-
- removeListenerOnError(SparkListenerInterface) - 类 中的方法org.apache.spark.scheduler.AsyncEventQueue
-
- removeMapOutput(int, BlockManagerId) - 类 中的方法org.apache.spark.ShuffleStatus
-
Remove the map output which was served by the specified block manager.
- removeOutputsByFilter(Function1<BlockManagerId, Object>) - 类 中的方法org.apache.spark.ShuffleStatus
-
Removes all shuffle outputs which satisfies the filter.
- removeOutputsOnExecutor(String) - 类 中的方法org.apache.spark.ShuffleStatus
-
Removes all map outputs associated with the specified executor.
- removeOutputsOnHost(String) - 类 中的方法org.apache.spark.ShuffleStatus
-
Removes all shuffle outputs associated with this host.
- removePartition(String) - 类 中的方法org.apache.spark.status.LiveRDD
-
- removePartition(LiveRDDPartition) - 类 中的方法org.apache.spark.status.RDDPartitionSeq
-
- RemoveRdd(int) - 类 的构造器org.apache.spark.storage.BlockManagerMessages.RemoveRdd
-
- RemoveRdd$() - 类 的构造器org.apache.spark.storage.BlockManagerMessages.RemoveRdd$
-
- removeReason() - 类 中的方法org.apache.spark.status.api.v1.ExecutorSummary
-
- removeReason() - 类 中的方法org.apache.spark.status.LiveExecutor
-
- removeSelfEdges() - 类 中的方法org.apache.spark.graphx.GraphOps
-
Remove self edges.
- RemoveShuffle(int) - 类 的构造器org.apache.spark.storage.BlockManagerMessages.RemoveShuffle
-
- RemoveShuffle$() - 类 的构造器org.apache.spark.storage.BlockManagerMessages.RemoveShuffle$
-
- removeShutdownDeleteDir(File) - 类 中的静态方法org.apache.spark.util.ShutdownHookManager
-
- removeShutdownHook(Object) - 类 中的静态方法org.apache.spark.util.ShutdownHookManager
-
Remove a previously installed shutdown hook.
- removeSparkListener(SparkListenerInterface) - 类 中的方法org.apache.spark.SparkContext
-
:: DeveloperApi ::
Deregister the listener from Spark's listener bus.
- removeTime() - 类 中的方法org.apache.spark.status.api.v1.ExecutorSummary
-
- removeTime() - 类 中的方法org.apache.spark.status.LiveExecutor
-
- RemoveWorker(String, String, String) - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RemoveWorker
-
- RemoveWorker$() - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RemoveWorker$
-
- rep(Function0<Parsers.Parser<T>>) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaParser
-
- rep1(Function0<Parsers.Parser<T>>) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaParser
-
- rep1(Function0<Parsers.Parser<T>>, Function0<Parsers.Parser<T>>) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaParser
-
- rep1sep(Function0<Parsers.Parser<T>>, Function0<Parsers.Parser<Object>>) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaParser
-
- repartition(int) - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
Return a new RDD that has exactly numPartitions partitions.
- repartition(int) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Return a new RDD that has exactly numPartitions partitions.
- repartition(int) - 类 中的方法org.apache.spark.api.java.JavaRDD
-
Return a new RDD that has exactly numPartitions partitions.
- repartition(int, Ordering<T>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- repartition(int, Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- repartition(int, Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- repartition(int, Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- repartition(int, Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- repartition(int, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- repartition(int, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- repartition(int, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- repartition(int, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- repartition(int, Ordering<T>) - 类 中的方法org.apache.spark.rdd.RDD
-
Return a new RDD that has exactly numPartitions partitions.
- repartition(int, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- repartition(int, Column...) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new Dataset partitioned by the given partitioning expressions into
numPartitions.
- repartition(Column...) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new Dataset partitioned by the given partitioning expressions, using
spark.sql.shuffle.partitions as number of partitions.
- repartition(int) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new Dataset that has exactly numPartitions partitions.
- repartition(int, Seq<Column>) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new Dataset partitioned by the given partitioning expressions into
numPartitions.
- repartition(Seq<Column>) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new Dataset partitioned by the given partitioning expressions, using
spark.sql.shuffle.partitions as number of partitions.
- repartition(int) - 类 中的方法org.apache.spark.streaming.api.java.JavaDStream
-
Return a new DStream with an increased or decreased level of parallelism.
- repartition(int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- repartition(int) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream with an increased or decreased level of parallelism.
- repartition(int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- repartition(int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- repartition(int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- repartition(int) - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Return a new DStream with an increased or decreased level of parallelism.
- repartition$default$2(int) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- repartition$default$2(int) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- repartition$default$2(int) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- repartition$default$2(int) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- repartition$default$2(int) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- repartition$default$2(int) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- repartition$default$2(int) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- repartition$default$2(int) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- repartition$default$2(int) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- repartition$default$2(int) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- repartitionAndSortWithinPartitions(Partitioner) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Repartition the RDD according to the given partitioner and, within each resulting partition,
sort records by their keys.
- repartitionAndSortWithinPartitions(Partitioner, Comparator<K>) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Repartition the RDD according to the given partitioner and, within each resulting partition,
sort records by their keys.
- repartitionAndSortWithinPartitions(Partitioner) - 类 中的方法org.apache.spark.rdd.OrderedRDDFunctions
-
Repartition the RDD according to the given partitioner and, within each resulting partition,
sort records by their keys.
- repartitionByRange(int, Column...) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new Dataset partitioned by the given partitioning expressions into
numPartitions.
- repartitionByRange(Column...) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new Dataset partitioned by the given partitioning expressions, using
spark.sql.shuffle.partitions as number of partitions.
- repartitionByRange(int, Seq<Column>) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new Dataset partitioned by the given partitioning expressions into
numPartitions.
- repartitionByRange(Seq<Column>) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new Dataset partitioned by the given partitioning expressions, using
spark.sql.shuffle.partitions as number of partitions.
- repeat(Column, int) - 类 中的静态方法org.apache.spark.sql.functions
-
Repeats a string column n times, and returns it as a new string column.
- replace(String, Map<T, T>) - 类 中的方法org.apache.spark.sql.DataFrameNaFunctions
-
Replaces values matching keys in replacement map with the corresponding values.
- replace(String[], Map<T, T>) - 类 中的方法org.apache.spark.sql.DataFrameNaFunctions
-
Replaces values matching keys in replacement map with the corresponding values.
- replace(String, Map<T, T>) - 类 中的方法org.apache.spark.sql.DataFrameNaFunctions
-
(Scala-specific) Replaces values matching keys in replacement map.
- replace(Seq<String>, Map<T, T>) - 类 中的方法org.apache.spark.sql.DataFrameNaFunctions
-
(Scala-specific) Replaces values matching keys in replacement map.
- replaceCharType(DataType) - 类 中的静态方法org.apache.spark.sql.types.HiveStringType
-
- replicas() - 类 中的方法org.apache.spark.storage.BlockManagerMessages.ReplicateBlock
-
- ReplicateBlock(BlockId, Seq<BlockManagerId>, int) - 类 的构造器org.apache.spark.storage.BlockManagerMessages.ReplicateBlock
-
- ReplicateBlock$() - 类 的构造器org.apache.spark.storage.BlockManagerMessages.ReplicateBlock$
-
- replicatedVertexView() - 类 中的方法org.apache.spark.graphx.impl.GraphImpl
-
- replication() - 类 中的方法org.apache.spark.storage.StorageLevel
-
- repN(int, Function0<Parsers.Parser<T>>) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaParser
-
- reportError(String, Throwable) - 类 中的方法org.apache.spark.streaming.receiver.Receiver
-
Report exceptions in receiving data.
- repr() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- repsep(Function0<Parsers.Parser<T>>, Function0<Parsers.Parser<Object>>) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaParser
-
- requestedTotal() - 类 中的方法org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RequestExecutors
-
- RequestExecutors(int, int, Map<String, Object>, Set<String>) - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RequestExecutors
-
- requestExecutors(int) - 类 中的方法org.apache.spark.SparkContext
-
:: DeveloperApi ::
Request an additional number of executors from the cluster manager.
- RequestExecutors$() - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RequestExecutors$
-
- requestTotalExecutors(int, int, Map<String, Object>) - 类 中的方法org.apache.spark.SparkContext
-
Update the cluster manager on our scheduling needs.
- requiredChildDistribution() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- requiredChildOrdering() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- res() - 类 中的方法org.apache.spark.mllib.optimization.NNLS.Workspace
-
- reservoirSampleAndCount(Iterator<T>, int, long, ClassTag<T>) - 类 中的静态方法org.apache.spark.util.random.SamplingUtils
-
Reservoir sampling implementation that also returns the input size.
- reset() - 类 中的静态方法org.apache.spark.metrics.source.HiveCatalogMetrics
-
Resets the values of all metrics to zero.
- reset() - 类 中的方法org.apache.spark.storage.BufferReleasingInputStream
-
- reset() - 类 中的方法org.apache.spark.util.AccumulatorV2
-
Resets this accumulator, which is zero value. i.e. call isZero must
return true.
- reset() - 类 中的方法org.apache.spark.util.CollectionAccumulator
-
- reset() - 类 中的方法org.apache.spark.util.DoubleAccumulator
-
- reset() - 类 中的方法org.apache.spark.util.LegacyAccumulatorWrapper
-
- reset() - 类 中的方法org.apache.spark.util.LongAccumulator
-
- resetMetrics() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- resetTerminated() - 类 中的方法org.apache.spark.sql.streaming.StreamingQueryManager
-
Forget about past terminated queries so that awaitAnyTermination() can be used again to
wait for new terminations.
- residualDegreeOfFreedom() - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegressionSummary
-
The residual degrees of freedom.
- residualDegreeOfFreedomNull() - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegressionSummary
-
The residual degrees of freedom for the null model.
- residuals() - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegressionSummary
-
Get the default residuals (deviance residuals) of the fitted model.
- residuals(String) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegressionSummary
-
Get the residuals of the fitted model by type.
- residuals() - 类 中的方法org.apache.spark.ml.regression.LinearRegressionSummary
-
Residuals (label - predicted value)
- resolve(StructType, Function2<String, String, Object>) - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- resolve(Seq<String>, Function2<String, String, Object>) - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- resolve(StructType, Function2<String, String, Object>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- resolve(Seq<String>, Function2<String, String, Object>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- resolve(StructType, Function2<String, String, Object>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- resolve(Seq<String>, Function2<String, String, Object>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- resolveChildren(Seq<String>, Function2<String, String, Object>) - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- resolveChildren(Seq<String>, Function2<String, String, Object>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- resolveChildren(Seq<String>, Function2<String, String, Object>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- resolved() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- resolved() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- resolved() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- ResolveHiveSerdeTable - org.apache.spark.sql.hive中的类
-
Determine the database, serde/format and schema of the Hive serde table, according to the storage
properties.
- ResolveHiveSerdeTable(SparkSession) - 类 的构造器org.apache.spark.sql.hive.ResolveHiveSerdeTable
-
- resolveQuoted(String, Function2<String, String, Object>) - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- resolveQuoted(String, Function2<String, String, Object>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- resolveQuoted(String, Function2<String, String, Object>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- resolveURI(String) - 类 中的静态方法org.apache.spark.util.Utils
-
Return a well-formed URI for the file described by a user input string.
- resolveURIs(String) - 类 中的静态方法org.apache.spark.util.Utils
-
Resolve a comma-separated list of paths.
- responder() - 类 中的方法org.apache.spark.ui.JettyUtils.ServletParams
-
- responseFromBackup(String) - 类 中的静态方法org.apache.spark.util.Utils
-
Return true if the response message is sent from a backup Master on standby.
- restart(String) - 类 中的方法org.apache.spark.streaming.receiver.Receiver
-
Restart the receiver.
- restart(String, Throwable) - 类 中的方法org.apache.spark.streaming.receiver.Receiver
-
Restart the receiver.
- restart(String, Throwable, int) - 类 中的方法org.apache.spark.streaming.receiver.Receiver
-
Restart the receiver.
- ResubmitFailedStages - org.apache.spark.scheduler中的类
-
- ResubmitFailedStages() - 类 的构造器org.apache.spark.scheduler.ResubmitFailedStages
-
- Resubmitted - org.apache.spark中的类
-
:: DeveloperApi ::
A org.apache.spark.scheduler.ShuffleMapTask that completed successfully earlier, but we
lost the executor before the stage completed.
- Resubmitted() - 类 的构造器org.apache.spark.Resubmitted
-
- result(Duration, CanAwait) - 类 中的方法org.apache.spark.ComplexFutureAction
-
- result(Duration, CanAwait) - 接口 中的方法org.apache.spark.FutureAction
-
Awaits and returns the result (of type T) of this action.
- result(Duration, CanAwait) - 类 中的方法org.apache.spark.SimpleFutureAction
-
- RESULT_SERIALIZATION_TIME() - 类 中的静态方法org.apache.spark.InternalAccumulator
-
- RESULT_SERIALIZATION_TIME() - 类 中的静态方法org.apache.spark.ui.jobs.TaskDetailsClassNames
-
- RESULT_SERIALIZATION_TIME() - 类 中的静态方法org.apache.spark.ui.ToolTips
-
- RESULT_SIZE() - 类 中的静态方法org.apache.spark.InternalAccumulator
-
- RESULT_SIZE() - 类 中的静态方法org.apache.spark.status.TaskIndexNames
-
- resultFetchStart() - 类 中的方法org.apache.spark.status.api.v1.TaskData
-
- resultSerializationTime() - 类 中的方法org.apache.spark.status.api.v1.TaskMetricDistributions
-
- resultSerializationTime() - 类 中的方法org.apache.spark.status.api.v1.TaskMetrics
-
- resultSetToObjectArray(ResultSet) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- resultSize() - 类 中的方法org.apache.spark.status.api.v1.TaskMetricDistributions
-
- resultSize() - 类 中的方法org.apache.spark.status.api.v1.TaskMetrics
-
- RetrieveLastAllocatedExecutorId$() - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RetrieveLastAllocatedExecutorId$
-
- RetrieveSparkAppConfig$() - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.RetrieveSparkAppConfig$
-
- retryWaitMs(SparkConf) - 类 中的静态方法org.apache.spark.util.RpcUtils
-
Returns the configured number of milliseconds to wait on each retry
- ReturnStatementFinder - org.apache.spark.util中的类
-
- ReturnStatementFinder() - 类 的构造器org.apache.spark.util.ReturnStatementFinder
-
- reverse() - 类 中的方法org.apache.spark.graphx.EdgeDirection
-
Reverse the direction of an edge.
- reverse() - 类 中的方法org.apache.spark.graphx.EdgeRDD
-
Reverse all the edges in this RDD.
- reverse() - 类 中的方法org.apache.spark.graphx.Graph
-
Reverses all edges in the graph.
- reverse() - 类 中的方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- reverse() - 类 中的方法org.apache.spark.graphx.impl.GraphImpl
-
- reverse(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Reverses the string column and returns it as a new string column.
- reverse() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- reverse() - 类 中的方法org.apache.spark.util.kvstore.KVStoreView
-
Reverses the order of iteration.
- reverseIterator() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- reverseMap(Function1<A, B>, CanBuildFrom<Repr, B, That>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- reverseRoutingTables() - 类 中的方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- reverseRoutingTables() - 类 中的方法org.apache.spark.graphx.VertexRDD
-
Returns a new
VertexRDD reflecting a reversal of all edge directions in the corresponding
EdgeRDD.
- ReviveOffers - org.apache.spark.scheduler.local中的类
-
- ReviveOffers() - 类 的构造器org.apache.spark.scheduler.local.ReviveOffers
-
- ReviveOffers$() - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.ReviveOffers$
-
- RFormula - org.apache.spark.ml.feature中的类
-
:: Experimental ::
Implements the transforms required for fitting a dataset against an R model formula.
- RFormula(String) - 类 的构造器org.apache.spark.ml.feature.RFormula
-
- RFormula() - 类 的构造器org.apache.spark.ml.feature.RFormula
-
- RFormulaModel - org.apache.spark.ml.feature中的类
-
:: Experimental ::
Model fitted by
RFormula.
- RFormulaParser - org.apache.spark.ml.feature中的类
-
Limited implementation of R formula parsing.
- RFormulaParser() - 类 的构造器org.apache.spark.ml.feature.RFormulaParser
-
- RidgeRegressionModel - org.apache.spark.mllib.regression中的类
-
Regression model trained using RidgeRegression.
- RidgeRegressionModel(Vector, double) - 类 的构造器org.apache.spark.mllib.regression.RidgeRegressionModel
-
- RidgeRegressionWithSGD - org.apache.spark.mllib.regression中的类
-
Train a regression model with L2-regularization using Stochastic Gradient Descent.
- RidgeRegressionWithSGD() - 类 的构造器org.apache.spark.mllib.regression.RidgeRegressionWithSGD
-
- right() - 类 中的方法org.apache.spark.sql.sources.And
-
- right() - 类 中的方法org.apache.spark.sql.sources.Or
-
- rightCategories() - 类 中的方法org.apache.spark.ml.tree.CategoricalSplit
-
Get sorted categories which split to the right
- rightChild() - 类 中的方法org.apache.spark.ml.tree.DecisionTreeModelReadWrite.NodeData
-
- rightChild() - 类 中的方法org.apache.spark.ml.tree.InternalNode
-
- rightChildIndex(int) - 类 中的静态方法org.apache.spark.mllib.tree.model.Node
-
Return the index of the right child of this node.
- rightImpurity() - 类 中的方法org.apache.spark.mllib.tree.model.InformationGainStats
-
- rightNode() - 类 中的方法org.apache.spark.mllib.tree.model.Node
-
- rightNodeId() - 类 中的方法org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.NodeData
-
- rightOuterJoin(JavaPairRDD<K, W>, Partitioner) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Perform a right outer join of this and other.
- rightOuterJoin(JavaPairRDD<K, W>) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Perform a right outer join of this and other.
- rightOuterJoin(JavaPairRDD<K, W>, int) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Perform a right outer join of this and other.
- rightOuterJoin(RDD<Tuple2<K, W>>, Partitioner) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Perform a right outer join of this and other.
- rightOuterJoin(RDD<Tuple2<K, W>>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Perform a right outer join of this and other.
- rightOuterJoin(RDD<Tuple2<K, W>>, int) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Perform a right outer join of this and other.
- rightOuterJoin(JavaPairDStream<K, W>) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying 'right outer join' between RDDs of this DStream and
other DStream.
- rightOuterJoin(JavaPairDStream<K, W>, int) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying 'right outer join' between RDDs of this DStream and
other DStream.
- rightOuterJoin(JavaPairDStream<K, W>, Partitioner) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Return a new DStream by applying 'right outer join' between RDDs of this DStream and
other DStream.
- rightOuterJoin(JavaPairDStream<K, W>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- rightOuterJoin(JavaPairDStream<K, W>, int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- rightOuterJoin(JavaPairDStream<K, W>, Partitioner) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- rightOuterJoin(JavaPairDStream<K, W>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- rightOuterJoin(JavaPairDStream<K, W>, int) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- rightOuterJoin(JavaPairDStream<K, W>, Partitioner) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- rightOuterJoin(DStream<Tuple2<K, W>>, ClassTag<W>) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying 'right outer join' between RDDs of this DStream and
other DStream.
- rightOuterJoin(DStream<Tuple2<K, W>>, int, ClassTag<W>) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying 'right outer join' between RDDs of this DStream and
other DStream.
- rightOuterJoin(DStream<Tuple2<K, W>>, Partitioner, ClassTag<W>) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Return a new DStream by applying 'right outer join' between RDDs of this DStream and
other DStream.
- rightPredict() - 类 中的方法org.apache.spark.mllib.tree.model.InformationGainStats
-
- rint(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the double value that is closest in value to the argument and
is equal to a mathematical integer.
- rint(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the double value that is closest in value to the argument and
is equal to a mathematical integer.
- rlike(String) - 类 中的方法org.apache.spark.sql.Column
-
SQL RLIKE expression (LIKE with Regex).
- RMATa() - 类 中的静态方法org.apache.spark.graphx.util.GraphGenerators
-
- RMATb() - 类 中的静态方法org.apache.spark.graphx.util.GraphGenerators
-
- RMATc() - 类 中的静态方法org.apache.spark.graphx.util.GraphGenerators
-
- RMATd() - 类 中的静态方法org.apache.spark.graphx.util.GraphGenerators
-
- rmatGraph(SparkContext, int, int) - 类 中的静态方法org.apache.spark.graphx.util.GraphGenerators
-
A random graph generator using the R-MAT model, proposed in
"R-MAT: A Recursive Model for Graph Mining" by Chakrabarti et al.
- rnd() - 类 中的方法org.apache.spark.rdd.DefaultPartitionCoalescer
-
- roc() - 接口 中的方法org.apache.spark.ml.classification.BinaryLogisticRegressionSummary
-
Returns the receiver operating characteristic (ROC) curve,
which is a Dataframe having two fields (FPR, TPR)
with (0.0, 0.0) prepended and (1.0, 1.0) appended to it.
- roc() - 类 中的方法org.apache.spark.mllib.evaluation.BinaryClassificationMetrics
-
Returns the receiver operating characteristic (ROC) curve,
which is an RDD of (false positive rate, true positive rate)
with (0.0, 0.0) prepended and (1.0, 1.0) appended to it.
- rollup(Column...) - 类 中的方法org.apache.spark.sql.Dataset
-
Create a multi-dimensional rollup for the current Dataset using the specified columns,
so we can run aggregation on them.
- rollup(String, String...) - 类 中的方法org.apache.spark.sql.Dataset
-
Create a multi-dimensional rollup for the current Dataset using the specified columns,
so we can run aggregation on them.
- rollup(Seq<Column>) - 类 中的方法org.apache.spark.sql.Dataset
-
Create a multi-dimensional rollup for the current Dataset using the specified columns,
so we can run aggregation on them.
- rollup(String, Seq<String>) - 类 中的方法org.apache.spark.sql.Dataset
-
Create a multi-dimensional rollup for the current Dataset using the specified columns,
so we can run aggregation on them.
- RollupType$() - 类 的构造器org.apache.spark.sql.RelationalGroupedDataset.RollupType$
-
- rootMeanSquaredError() - 类 中的方法org.apache.spark.ml.regression.LinearRegressionSummary
-
Returns the root mean squared error, which is defined as the square root of
the mean squared error.
- rootMeanSquaredError() - 类 中的方法org.apache.spark.mllib.evaluation.RegressionMetrics
-
Returns the root mean squared error, which is defined as the square root of
the mean squared error.
- rootNode() - 类 中的方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- rootNode() - 类 中的方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- round(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the value of the column e rounded to 0 decimal places with HALF_UP round mode.
- round(Column, int) - 类 中的静态方法org.apache.spark.sql.functions
-
Round the value of e to scale decimal places with HALF_UP round mode
if scale is greater than or equal to 0 or at integral part when scale is less than 0.
- ROUND_CEILING() - 类 中的静态方法org.apache.spark.sql.types.Decimal
-
- ROUND_FLOOR() - 类 中的静态方法org.apache.spark.sql.types.Decimal
-
- ROUND_HALF_EVEN() - 类 中的静态方法org.apache.spark.sql.types.Decimal
-
- ROUND_HALF_UP() - 类 中的静态方法org.apache.spark.sql.types.Decimal
-
- ROW() - 类 中的静态方法org.apache.spark.api.r.SerializationFormats
-
- Row - org.apache.spark.sql中的接口
-
Represents one row of output from a relational operator.
- row_number() - 类 中的静态方法org.apache.spark.sql.functions
-
Window function: returns a sequential number starting at 1 within a window partition.
- RowFactory - org.apache.spark.sql中的类
-
A factory class used to construct
Row objects.
- RowFactory() - 类 的构造器org.apache.spark.sql.RowFactory
-
- rowIndices() - 类 中的方法org.apache.spark.ml.linalg.SparseMatrix
-
- rowIndices() - 类 中的方法org.apache.spark.mllib.linalg.SparseMatrix
-
- rowIter() - 类 中的静态方法org.apache.spark.ml.linalg.DenseMatrix
-
- rowIter() - 接口 中的方法org.apache.spark.ml.linalg.Matrix
-
Returns an iterator of row vectors.
- rowIter() - 类 中的静态方法org.apache.spark.ml.linalg.SparseMatrix
-
- rowIter() - 类 中的静态方法org.apache.spark.mllib.linalg.DenseMatrix
-
- rowIter() - 接口 中的方法org.apache.spark.mllib.linalg.Matrix
-
Returns an iterator of row vectors.
- rowIter() - 类 中的静态方法org.apache.spark.mllib.linalg.SparseMatrix
-
- rowIterator() - 类 中的方法org.apache.spark.sql.vectorized.ColumnarBatch
-
Returns an iterator over the rows in this batch.
- RowMatrix - org.apache.spark.mllib.linalg.distributed中的类
-
Represents a row-oriented distributed Matrix with no meaningful row indices.
- RowMatrix(RDD<Vector>, long, int) - 类 的构造器org.apache.spark.mllib.linalg.distributed.RowMatrix
-
- RowMatrix(RDD<Vector>) - 类 的构造器org.apache.spark.mllib.linalg.distributed.RowMatrix
-
Alternative constructor leaving matrix dimensions to be determined automatically.
- rows() - 类 中的方法org.apache.spark.mllib.linalg.distributed.IndexedRowMatrix
-
- rows() - 类 中的方法org.apache.spark.mllib.linalg.distributed.RowMatrix
-
- rowsBetween(long, long) - 类 中的静态方法org.apache.spark.sql.expressions.Window
-
Creates a
WindowSpec with the frame boundaries defined,
from
start (inclusive) to
end (inclusive).
- rowsBetween(long, long) - 类 中的方法org.apache.spark.sql.expressions.WindowSpec
-
Defines the frame boundaries, from start (inclusive) to end (inclusive).
- rowsPerBlock() - 类 中的方法org.apache.spark.mllib.linalg.distributed.BlockMatrix
-
- rPackages() - 类 中的静态方法org.apache.spark.api.r.RUtils
-
- rpad(Column, int, String) - 类 中的静态方法org.apache.spark.sql.functions
-
Right-pad the string column with pad to a length of len.
- RpcUtils - org.apache.spark.util中的类
-
- RpcUtils() - 类 的构造器org.apache.spark.util.RpcUtils
-
- RRDD<T> - org.apache.spark.api.r中的类
-
An RDD that stores serialized R objects as Array[Byte].
- RRDD(RDD<T>, byte[], String, String, byte[], Object[], ClassTag<T>) - 类 的构造器org.apache.spark.api.r.RRDD
-
- RRunnerModes - org.apache.spark.api.r中的类
-
- RRunnerModes() - 类 的构造器org.apache.spark.api.r.RRunnerModes
-
- rtrim(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Trim the spaces from right end for the specified string value.
- rtrim(Column, String) - 类 中的静态方法org.apache.spark.sql.functions
-
Trim the specified character string from right end for the specified string column.
- ruleName() - 类 中的静态方法org.apache.spark.sql.hive.HiveAnalysis
-
- ruleName() - 类 中的静态方法org.apache.spark.sql.hive.RelationConversions
-
- run(Graph<VD, ED>, int, ClassTag<VD>, ClassTag<ED>) - 类 中的静态方法org.apache.spark.graphx.lib.ConnectedComponents
-
Compute the connected component membership of each vertex and return a graph with the vertex
value containing the lowest vertex id in the connected component containing that vertex.
- run(Graph<VD, ED>, ClassTag<VD>, ClassTag<ED>) - 类 中的静态方法org.apache.spark.graphx.lib.ConnectedComponents
-
Compute the connected component membership of each vertex and return a graph with the vertex
value containing the lowest vertex id in the connected component containing that vertex.
- run(Graph<VD, ED>, int, ClassTag<ED>) - 类 中的静态方法org.apache.spark.graphx.lib.LabelPropagation
-
Run static Label Propagation for detecting communities in networks.
- run(Graph<VD, ED>, int, double, ClassTag<VD>, ClassTag<ED>) - 类 中的静态方法org.apache.spark.graphx.lib.PageRank
-
Run PageRank for a fixed number of iterations returning a graph
with vertex attributes containing the PageRank and edge
attributes the normalized edge weight.
- run(Graph<VD, ED>, Seq<Object>, ClassTag<ED>) - 类 中的静态方法org.apache.spark.graphx.lib.ShortestPaths
-
Computes shortest paths to the given set of landmark vertices.
- run(Graph<VD, ED>, int, ClassTag<VD>, ClassTag<ED>) - 类 中的静态方法org.apache.spark.graphx.lib.StronglyConnectedComponents
-
Compute the strongly connected component (SCC) of each vertex and return a graph with the
vertex value containing the lowest vertex id in the SCC containing that vertex.
- run(RDD<Edge<Object>>, SVDPlusPlus.Conf) - 类 中的静态方法org.apache.spark.graphx.lib.SVDPlusPlus
-
Implement SVD++ based on "Factorization Meets the Neighborhood:
a Multifaceted Collaborative Filtering Model",
available at
here.
- run(Graph<VD, ED>, ClassTag<VD>, ClassTag<ED>) - 类 中的静态方法org.apache.spark.graphx.lib.TriangleCount
-
- run(RDD<LabeledPoint>, BoostingStrategy, long, String) - 类 中的静态方法org.apache.spark.ml.tree.impl.GradientBoostedTrees
-
Method to train a gradient boosting model
- run(RDD<LabeledPoint>, Strategy, int, String, long, Option<<any>>, Option<String>) - 类 中的静态方法org.apache.spark.ml.tree.impl.RandomForest
-
Train a random forest.
- run(RDD<LabeledPoint>) - 类 中的方法org.apache.spark.mllib.classification.LogisticRegressionWithLBFGS
-
Run Logistic Regression with the configured parameters on an input RDD
of LabeledPoint entries.
- run(RDD<LabeledPoint>, Vector) - 类 中的方法org.apache.spark.mllib.classification.LogisticRegressionWithLBFGS
-
Run Logistic Regression with the configured parameters on an input RDD
of LabeledPoint entries starting from the initial weights provided.
- run(RDD<LabeledPoint>) - 类 中的静态方法org.apache.spark.mllib.classification.LogisticRegressionWithSGD
-
- run(RDD<LabeledPoint>, Vector) - 类 中的静态方法org.apache.spark.mllib.classification.LogisticRegressionWithSGD
-
- run(RDD<LabeledPoint>) - 类 中的方法org.apache.spark.mllib.classification.NaiveBayes
-
Run the algorithm with the configured parameters on an input RDD of LabeledPoint entries.
- run(RDD<LabeledPoint>) - 类 中的静态方法org.apache.spark.mllib.classification.SVMWithSGD
-
- run(RDD<LabeledPoint>, Vector) - 类 中的静态方法org.apache.spark.mllib.classification.SVMWithSGD
-
- run(RDD<Vector>) - 类 中的方法org.apache.spark.mllib.clustering.BisectingKMeans
-
Runs the bisecting k-means algorithm.
- run(JavaRDD<Vector>) - 类 中的方法org.apache.spark.mllib.clustering.BisectingKMeans
-
Java-friendly version of run().
- run(RDD<Vector>) - 类 中的方法org.apache.spark.mllib.clustering.GaussianMixture
-
Perform expectation maximization
- run(JavaRDD<Vector>) - 类 中的方法org.apache.spark.mllib.clustering.GaussianMixture
-
Java-friendly version of run()
- run(RDD<Vector>) - 类 中的方法org.apache.spark.mllib.clustering.KMeans
-
Train a K-means model on the given set of points; data should be cached for high
performance, because this is an iterative algorithm.
- run(RDD<Tuple2<Object, Vector>>) - 类 中的方法org.apache.spark.mllib.clustering.LDA
-
Learn an LDA model using the given dataset.
- run(JavaPairRDD<Long, Vector>) - 类 中的方法org.apache.spark.mllib.clustering.LDA
-
Java-friendly version of run()
- run(Graph<Object, Object>) - 类 中的方法org.apache.spark.mllib.clustering.PowerIterationClustering
-
Run the PIC algorithm on Graph.
- run(RDD<Tuple3<Object, Object, Object>>) - 类 中的方法org.apache.spark.mllib.clustering.PowerIterationClustering
-
Run the PIC algorithm.
- run(JavaRDD<Tuple3<Long, Long, Double>>) - 类 中的方法org.apache.spark.mllib.clustering.PowerIterationClustering
-
A Java-friendly version of PowerIterationClustering.run.
- run(RDD<FPGrowth.FreqItemset<Item>>, ClassTag<Item>) - 类 中的方法org.apache.spark.mllib.fpm.AssociationRules
-
Computes the association rules with confidence above minConfidence.
- run(JavaRDD<FPGrowth.FreqItemset<Item>>) - 类 中的方法org.apache.spark.mllib.fpm.AssociationRules
-
Java-friendly version of run.
- run(RDD<Object>, ClassTag<Item>) - 类 中的方法org.apache.spark.mllib.fpm.FPGrowth
-
Computes an FP-Growth model that contains frequent itemsets.
- run(JavaRDD<Basket>) - 类 中的方法org.apache.spark.mllib.fpm.FPGrowth
-
Java-friendly version of run.
- run(RDD<Object[]>, ClassTag<Item>) - 类 中的方法org.apache.spark.mllib.fpm.PrefixSpan
-
Finds the complete set of frequent sequential patterns in the input sequences of itemsets.
- run(JavaRDD<Sequence>) - 类 中的方法org.apache.spark.mllib.fpm.PrefixSpan
-
A Java-friendly version of
run() that reads sequences from a
JavaRDD and returns
frequent sequences in a
PrefixSpanModel.
- run(RDD<Rating>) - 类 中的方法org.apache.spark.mllib.recommendation.ALS
-
Run ALS with the configured parameters on an input RDD of
Rating objects.
- run(JavaRDD<Rating>) - 类 中的方法org.apache.spark.mllib.recommendation.ALS
-
Java-friendly version of ALS.run.
- run(RDD<LabeledPoint>) - 类 中的方法org.apache.spark.mllib.regression.GeneralizedLinearAlgorithm
-
Run the algorithm with the configured parameters on an input
RDD of LabeledPoint entries.
- run(RDD<LabeledPoint>, Vector) - 类 中的方法org.apache.spark.mllib.regression.GeneralizedLinearAlgorithm
-
Run the algorithm with the configured parameters on an input RDD
of LabeledPoint entries starting from the initial weights provided.
- run(RDD<Tuple3<Object, Object, Object>>) - 类 中的方法org.apache.spark.mllib.regression.IsotonicRegression
-
Run IsotonicRegression algorithm to obtain isotonic regression model.
- run(JavaRDD<Tuple3<Double, Double, Double>>) - 类 中的方法org.apache.spark.mllib.regression.IsotonicRegression
-
Run pool adjacent violators algorithm to obtain isotonic regression model.
- run(RDD<LabeledPoint>) - 类 中的静态方法org.apache.spark.mllib.regression.LassoWithSGD
-
- run(RDD<LabeledPoint>, Vector) - 类 中的静态方法org.apache.spark.mllib.regression.LassoWithSGD
-
- run(RDD<LabeledPoint>) - 类 中的静态方法org.apache.spark.mllib.regression.LinearRegressionWithSGD
-
- run(RDD<LabeledPoint>, Vector) - 类 中的静态方法org.apache.spark.mllib.regression.LinearRegressionWithSGD
-
- run(RDD<LabeledPoint>) - 类 中的静态方法org.apache.spark.mllib.regression.RidgeRegressionWithSGD
-
- run(RDD<LabeledPoint>, Vector) - 类 中的静态方法org.apache.spark.mllib.regression.RidgeRegressionWithSGD
-
- run(RDD<LabeledPoint>) - 类 中的方法org.apache.spark.mllib.tree.DecisionTree
-
Method to train a decision tree model over an RDD
- run(RDD<LabeledPoint>) - 类 中的方法org.apache.spark.mllib.tree.GradientBoostedTrees
-
Method to train a gradient boosting model
- run(JavaRDD<LabeledPoint>) - 类 中的方法org.apache.spark.mllib.tree.GradientBoostedTrees
-
Java-friendly API for org.apache.spark.mllib.tree.GradientBoostedTrees.run.
- run(RDD<LabeledPoint>) - 类 中的方法org.apache.spark.mllib.tree.RandomForest
-
Method to train a decision tree model over an RDD
- run(SparkSession, SparkPlan) - 类 中的方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- run(SparkSession, SparkPlan) - 类 中的方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- run(SparkSession, SparkPlan) - 类 中的方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
Inserts all the rows in the table into Hive.
- run() - 类 中的方法org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread
-
- run() - 类 中的方法org.apache.spark.util.SparkShutdownHook
-
- runApproximateJob(RDD<T>, Function2<TaskContext, Iterator<T>, U>, <any>, long) - 类 中的方法org.apache.spark.SparkContext
-
:: DeveloperApi ::
Run a job that can return approximate results.
- runId() - 接口 中的方法org.apache.spark.sql.streaming.StreamingQuery
-
Returns the unique id of this run of the query.
- runId() - 类 中的方法org.apache.spark.sql.streaming.StreamingQueryListener.QueryStartedEvent
-
- runId() - 类 中的方法org.apache.spark.sql.streaming.StreamingQueryListener.QueryTerminatedEvent
-
- runId() - 类 中的方法org.apache.spark.sql.streaming.StreamingQueryProgress
-
- runInNewThread(String, boolean, Function0<T>) - 类 中的静态方法org.apache.spark.util.ThreadUtils
-
Run a piece of code in a new thread and return the result.
- runJob(RDD<T>, Function2<TaskContext, Iterator<T>, U>, Seq<Object>, Function2<Object, U, BoxedUnit>, ClassTag<U>) - 类 中的方法org.apache.spark.SparkContext
-
Run a function on a given set of partitions in an RDD and pass the results to the given
handler function.
- runJob(RDD<T>, Function2<TaskContext, Iterator<T>, U>, Seq<Object>, ClassTag<U>) - 类 中的方法org.apache.spark.SparkContext
-
Run a function on a given set of partitions in an RDD and return the results as an array.
- runJob(RDD<T>, Function1<Iterator<T>, U>, Seq<Object>, ClassTag<U>) - 类 中的方法org.apache.spark.SparkContext
-
Run a function on a given set of partitions in an RDD and return the results as an array.
- runJob(RDD<T>, Function2<TaskContext, Iterator<T>, U>, ClassTag<U>) - 类 中的方法org.apache.spark.SparkContext
-
Run a job on all partitions in an RDD and return the results in an array.
- runJob(RDD<T>, Function1<Iterator<T>, U>, ClassTag<U>) - 类 中的方法org.apache.spark.SparkContext
-
Run a job on all partitions in an RDD and return the results in an array.
- runJob(RDD<T>, Function2<TaskContext, Iterator<T>, U>, Function2<Object, U, BoxedUnit>, ClassTag<U>) - 类 中的方法org.apache.spark.SparkContext
-
Run a job on all partitions in an RDD and pass the results to a handler function.
- runJob(RDD<T>, Function1<Iterator<T>, U>, Function2<Object, U, BoxedUnit>, ClassTag<U>) - 类 中的方法org.apache.spark.SparkContext
-
Run a job on all partitions in an RDD and pass the results to a handler function.
- runLBFGS(RDD<Tuple2<Object, Vector>>, Gradient, Updater, int, double, int, double, Vector) - 类 中的静态方法org.apache.spark.mllib.optimization.LBFGS
-
Run Limited-memory BFGS (L-BFGS) in parallel.
- runMiniBatchSGD(RDD<Tuple2<Object, Vector>>, Gradient, Updater, double, int, double, double, Vector, double) - 类 中的静态方法org.apache.spark.mllib.optimization.GradientDescent
-
Run stochastic gradient descent (SGD) in parallel using mini batches.
- runMiniBatchSGD(RDD<Tuple2<Object, Vector>>, Gradient, Updater, double, int, double, double, Vector) - 类 中的静态方法org.apache.spark.mllib.optimization.GradientDescent
-
Alias of runMiniBatchSGD with convergenceTol set to default value of 0.001.
- running() - 类 中的方法org.apache.spark.scheduler.TaskInfo
-
- RUNNING() - 类 中的静态方法org.apache.spark.TaskState
-
- runParallelPersonalizedPageRank(Graph<VD, ED>, int, double, long[], ClassTag<VD>, ClassTag<ED>) - 类 中的静态方法org.apache.spark.graphx.lib.PageRank
-
Run Personalized PageRank for a fixed number of iterations, for a
set of starting nodes in parallel.
- runPreCanonicalized(Graph<VD, ED>, ClassTag<VD>, ClassTag<ED>) - 类 中的静态方法org.apache.spark.graphx.lib.TriangleCount
-
- runtime() - 类 中的方法org.apache.spark.status.api.v1.ApplicationEnvironmentInfo
-
- RuntimeConfig - org.apache.spark.sql中的类
-
Runtime configuration interface for Spark.
- RuntimeInfo - org.apache.spark.status.api.v1中的类
-
- RuntimePercentage - org.apache.spark.scheduler中的类
-
- RuntimePercentage(double, Option<Object>, double) - 类 的构造器org.apache.spark.scheduler.RuntimePercentage
-
- runUntilConvergence(Graph<VD, ED>, double, double, ClassTag<VD>, ClassTag<ED>) - 类 中的静态方法org.apache.spark.graphx.lib.PageRank
-
Run a dynamic version of PageRank returning a graph with vertex attributes containing the
PageRank and edge attributes containing the normalized edge weight.
- runUntilConvergenceWithOptions(Graph<VD, ED>, double, double, Option<Object>, ClassTag<VD>, ClassTag<ED>) - 类 中的静态方法org.apache.spark.graphx.lib.PageRank
-
Run a dynamic version of PageRank returning a graph with vertex attributes containing the
PageRank and edge attributes containing the normalized edge weight.
- runWith(Function1<B, U>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- runWithOptions(Graph<VD, ED>, int, double, Option<Object>, ClassTag<VD>, ClassTag<ED>) - 类 中的静态方法org.apache.spark.graphx.lib.PageRank
-
Run PageRank for a fixed number of iterations returning a graph
with vertex attributes containing the PageRank and edge
attributes the normalized edge weight.
- runWithValidation(RDD<LabeledPoint>, RDD<LabeledPoint>, BoostingStrategy, long, String) - 类 中的静态方法org.apache.spark.ml.tree.impl.GradientBoostedTrees
-
Method to validate a gradient boosting model
- runWithValidation(RDD<LabeledPoint>, RDD<LabeledPoint>) - 类 中的方法org.apache.spark.mllib.tree.GradientBoostedTrees
-
Method to validate a gradient boosting model
- runWithValidation(JavaRDD<LabeledPoint>, JavaRDD<LabeledPoint>) - 类 中的方法org.apache.spark.mllib.tree.GradientBoostedTrees
-
Java-friendly API for org.apache.spark.mllib.tree.GradientBoostedTrees.runWithValidation.
- RUtils - org.apache.spark.api.r中的类
-
- RUtils() - 类 的构造器org.apache.spark.api.r.RUtils
-
- RWrappers - org.apache.spark.ml.r中的类
-
This is the Scala stub of SparkR read.ml.
- RWrappers() - 类 的构造器org.apache.spark.ml.r.RWrappers
-
- RWrapperUtils - org.apache.spark.ml.r中的类
-
- RWrapperUtils() - 类 的构造器org.apache.spark.ml.r.RWrapperUtils
-
- s() - 类 中的方法org.apache.spark.mllib.linalg.SingularValueDecomposition
-
- sameElements(GenIterable<B>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- sameResult(PlanType) - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- sameResult(PlanType) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- sameResult(PlanType) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- sameResult(PlanType) - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- sameThread() - 类 中的静态方法org.apache.spark.util.ThreadUtils
-
An ExecutionContextExecutor that runs each task in the thread that invokes execute/submit.
- sample(boolean, Double) - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
Return a sampled subset of this RDD.
- sample(boolean, Double, long) - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
Return a sampled subset of this RDD.
- sample(boolean, double) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Return a sampled subset of this RDD.
- sample(boolean, double, long) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Return a sampled subset of this RDD.
- sample(boolean, double) - 类 中的方法org.apache.spark.api.java.JavaRDD
-
Return a sampled subset of this RDD with a random seed.
- sample(boolean, double, long) - 类 中的方法org.apache.spark.api.java.JavaRDD
-
Return a sampled subset of this RDD, with a user-supplied seed.
- sample(boolean, double, long) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- sample(boolean, double, long) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- sample(boolean, double, long) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- sample(boolean, double, long) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- sample(boolean, double, long) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- sample(boolean, double, long) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- sample(boolean, double, long) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- sample(boolean, double, long) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- sample(boolean, double, long) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- sample(boolean, double, long) - 类 中的方法org.apache.spark.rdd.RDD
-
Return a sampled subset of this RDD.
- sample(boolean, double, long) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- sample(double, long) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new
Dataset by sampling a fraction of rows (without replacement),
using a user-supplied seed.
- sample(double) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new
Dataset by sampling a fraction of rows (without replacement),
using a random seed.
- sample(boolean, double, long) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new
Dataset by sampling a fraction of rows, using a user-supplied seed.
- sample(boolean, double) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new
Dataset by sampling a fraction of rows, using a random seed.
- sample() - 类 中的方法org.apache.spark.util.random.BernoulliCellSampler
-
- sample() - 类 中的方法org.apache.spark.util.random.BernoulliSampler
-
- sample() - 类 中的方法org.apache.spark.util.random.PoissonSampler
-
- sample(Iterator<T>) - 类 中的方法org.apache.spark.util.random.PoissonSampler
-
- sample(Iterator<T>) - 接口 中的方法org.apache.spark.util.random.RandomSampler
-
take a random sample
- sample() - 接口 中的方法org.apache.spark.util.random.RandomSampler
-
Whether to sample the next item or not.
- sample$default$3() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- sample$default$3() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- sample$default$3() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- sample$default$3() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- sample$default$3() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- sample$default$3() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- sample$default$3() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- sample$default$3() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- sample$default$3() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- sample$default$3() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- sampleBy(String, Map<T, Object>, long) - 类 中的方法org.apache.spark.sql.DataFrameStatFunctions
-
Returns a stratified sample without replacement based on the fraction given on each stratum.
- sampleBy(String, Map<T, Double>, long) - 类 中的方法org.apache.spark.sql.DataFrameStatFunctions
-
Returns a stratified sample without replacement based on the fraction given on each stratum.
- sampleByKey(boolean, Map<K, Double>, long) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Return a subset of this RDD sampled by key (via stratified sampling).
- sampleByKey(boolean, Map<K, Double>) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Return a subset of this RDD sampled by key (via stratified sampling).
- sampleByKey(boolean, Map<K, Object>, long) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Return a subset of this RDD sampled by key (via stratified sampling).
- sampleByKeyExact(boolean, Map<K, Double>, long) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Return a subset of this RDD sampled by key (via stratified sampling) containing exactly
math.ceil(numItems * samplingRate) for each stratum (group of pairs with the same key).
- sampleByKeyExact(boolean, Map<K, Double>) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Return a subset of this RDD sampled by key (via stratified sampling) containing exactly
math.ceil(numItems * samplingRate) for each stratum (group of pairs with the same key).
- sampleByKeyExact(boolean, Map<K, Object>, long) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Return a subset of this RDD sampled by key (via stratified sampling) containing exactly
math.ceil(numItems * samplingRate) for each stratum (group of pairs with the same key).
- SamplePathFilter - org.apache.spark.ml.image中的类
-
Filter that allows loading a fraction of HDFS files.
- SamplePathFilter() - 类 的构造器org.apache.spark.ml.image.SamplePathFilter
-
- samplePointsPerPartitionHint() - 类 中的方法org.apache.spark.RangePartitioner
-
- sampleRatio() - 类 中的方法org.apache.spark.ml.image.SamplePathFilter
-
- sampleStdev() - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
Compute the sample standard deviation of this RDD's elements (which corrects for bias in
estimating the standard deviation by dividing by N-1 instead of N).
- sampleStdev() - 类 中的方法org.apache.spark.rdd.DoubleRDDFunctions
-
Compute the sample standard deviation of this RDD's elements (which corrects for bias in
estimating the standard deviation by dividing by N-1 instead of N).
- sampleStdev() - 类 中的方法org.apache.spark.util.StatCounter
-
Return the sample standard deviation of the values, which corrects for bias in estimating the
variance by dividing by N-1 instead of N.
- sampleVariance() - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
Compute the sample variance of this RDD's elements (which corrects for bias in
estimating the standard variance by dividing by N-1 instead of N).
- sampleVariance() - 类 中的方法org.apache.spark.rdd.DoubleRDDFunctions
-
Compute the sample variance of this RDD's elements (which corrects for bias in
estimating the variance by dividing by N-1 instead of N).
- sampleVariance() - 类 中的方法org.apache.spark.util.StatCounter
-
Return the sample variance, which corrects for bias in estimating the variance by dividing
by N-1 instead of N.
- SamplingUtils - org.apache.spark.util.random中的类
-
- SamplingUtils() - 类 的构造器org.apache.spark.util.random.SamplingUtils
-
- satisfy(Distribution) - 接口 中的方法org.apache.spark.sql.sources.v2.reader.partitioning.Partitioning
-
Returns true if this partitioning can satisfy the given distribution, which means Spark does
not need to shuffle the output data of this data source for some certain operations.
- save(String) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- save(String) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- save(String) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- save(String) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- save(String) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- save(String) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- save(String) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.classification.OneVsRest
-
- save(String) - 类 中的静态方法org.apache.spark.ml.classification.OneVsRestModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- save(String) - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeans
-
- save(String) - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixture
-
- save(String) - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.clustering.KMeans
-
- save(String) - 类 中的静态方法org.apache.spark.ml.clustering.KMeansModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- save(String) - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
-
- save(String) - 类 中的静态方法org.apache.spark.ml.evaluation.ClusteringEvaluator
-
- save(String) - 类 中的静态方法org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- save(String) - 类 中的静态方法org.apache.spark.ml.evaluation.RegressionEvaluator
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.Binarizer
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.Bucketizer
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelector
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.ColumnPruner
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizer
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizerModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.DCT
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.ElementwiseProduct
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.FeatureHasher
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.HashingTF
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.IDF
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.IDFModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.Imputer
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.ImputerModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.IndexToString
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.Interaction
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.MaxAbsScaler
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.MaxAbsScalerModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSH
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSHModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScaler
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScalerModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.NGram
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.Normalizer
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoder
-
已过时。
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderEstimator
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.PCA
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.PCAModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.PolynomialExpansion
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.QuantileDiscretizer
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.RegexTokenizer
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.RFormula
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.SQLTransformer
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.StandardScaler
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.StandardScalerModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.StopWordsRemover
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.StringIndexer
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.StringIndexerModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.Tokenizer
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.VectorAssembler
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.VectorAttributeRewriter
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexer
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexerModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.VectorSizeHint
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.VectorSlicer
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.Word2Vec
-
- save(String) - 类 中的静态方法org.apache.spark.ml.feature.Word2VecModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowth
-
- save(String) - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowthModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.Pipeline
-
- save(String) - 类 中的静态方法org.apache.spark.ml.PipelineModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- save(String) - 类 中的静态方法org.apache.spark.ml.recommendation.ALSModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- save(String) - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- save(String) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- save(String) - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- save(String) - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegression
-
- save(String) - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegressionModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- save(String) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- save(String) - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidator
-
- save(String) - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidatorModel
-
- save(String) - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplit
-
- save(String) - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplitModel
-
- save(String) - 接口 中的方法org.apache.spark.ml.util.MLWritable
-
Saves this ML instance to the input path, a shortcut of write.save(path).
- save(String) - 类 中的方法org.apache.spark.ml.util.MLWriter
-
Saves the ML instances to the input path.
- save(SparkContext, String, String, int, int, Vector, double, Option<Object>) - 类 中的方法org.apache.spark.mllib.classification.impl.GLMClassificationModel.SaveLoadV1_0$
-
Helper method for saving GLM classification model metadata and data.
- save(SparkContext, String) - 类 中的方法org.apache.spark.mllib.classification.LogisticRegressionModel
-
- save(SparkContext, String) - 类 中的方法org.apache.spark.mllib.classification.NaiveBayesModel
-
- save(SparkContext, String, org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV1_0.Data) - 类 中的方法org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV1_0$
-
- save(SparkContext, String, org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV2_0.Data) - 类 中的方法org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV2_0$
-
- save(SparkContext, String) - 类 中的方法org.apache.spark.mllib.classification.SVMModel
-
- save(SparkContext, String) - 类 中的方法org.apache.spark.mllib.clustering.BisectingKMeansModel
-
- save(SparkContext, BisectingKMeansModel, String) - 类 中的方法org.apache.spark.mllib.clustering.BisectingKMeansModel.SaveLoadV1_0$
-
- save(SparkContext, String) - 类 中的方法org.apache.spark.mllib.clustering.DistributedLDAModel
-
- save(SparkContext, String) - 类 中的方法org.apache.spark.mllib.clustering.GaussianMixtureModel
-
- save(SparkContext, String) - 类 中的方法org.apache.spark.mllib.clustering.KMeansModel
-
- save(SparkContext, KMeansModel, String) - 类 中的方法org.apache.spark.mllib.clustering.KMeansModel.SaveLoadV1_0$
-
- save(SparkContext, String) - 类 中的方法org.apache.spark.mllib.clustering.LocalLDAModel
-
- save(SparkContext, String) - 类 中的方法org.apache.spark.mllib.clustering.PowerIterationClusteringModel
-
- save(SparkContext, PowerIterationClusteringModel, String) - 类 中的方法org.apache.spark.mllib.clustering.PowerIterationClusteringModel.SaveLoadV1_0$
-
- save(SparkContext, String) - 类 中的方法org.apache.spark.mllib.feature.ChiSqSelectorModel
-
- save(SparkContext, ChiSqSelectorModel, String) - 类 中的方法org.apache.spark.mllib.feature.ChiSqSelectorModel.SaveLoadV1_0$
-
- save(SparkContext, String) - 类 中的方法org.apache.spark.mllib.feature.Word2VecModel
-
- save(SparkContext, String) - 类 中的方法org.apache.spark.mllib.fpm.FPGrowthModel
-
Save this model to the given path.
- save(FPGrowthModel<?>, String) - 类 中的方法org.apache.spark.mllib.fpm.FPGrowthModel.SaveLoadV1_0$
-
- save(SparkContext, String) - 类 中的方法org.apache.spark.mllib.fpm.PrefixSpanModel
-
Save this model to the given path.
- save(PrefixSpanModel<?>, String) - 类 中的方法org.apache.spark.mllib.fpm.PrefixSpanModel.SaveLoadV1_0$
-
- save(SparkContext, String) - 类 中的方法org.apache.spark.mllib.recommendation.MatrixFactorizationModel
-
Save this model to the given path.
- save(MatrixFactorizationModel, String) - 类 中的方法org.apache.spark.mllib.recommendation.MatrixFactorizationModel.SaveLoadV1_0$
-
Saves a
MatrixFactorizationModel, where user features are saved under
data/users and
product features are saved under
data/products.
- save(SparkContext, String, String, Vector, double) - 类 中的方法org.apache.spark.mllib.regression.impl.GLMRegressionModel.SaveLoadV1_0$
-
Helper method for saving GLM regression model metadata and data.
- save(SparkContext, String) - 类 中的方法org.apache.spark.mllib.regression.IsotonicRegressionModel
-
- save(SparkContext, String) - 类 中的方法org.apache.spark.mllib.regression.LassoModel
-
- save(SparkContext, String) - 类 中的方法org.apache.spark.mllib.regression.LinearRegressionModel
-
- save(SparkContext, String) - 类 中的方法org.apache.spark.mllib.regression.RidgeRegressionModel
-
- save(SparkContext, String) - 类 中的方法org.apache.spark.mllib.tree.model.DecisionTreeModel
-
- save(SparkContext, String, DecisionTreeModel) - 类 中的方法org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$
-
- save(SparkContext, String) - 类 中的方法org.apache.spark.mllib.tree.model.GradientBoostedTreesModel
-
- save(SparkContext, String) - 类 中的方法org.apache.spark.mllib.tree.model.RandomForestModel
-
- save(SparkContext, String) - 接口 中的方法org.apache.spark.mllib.util.Saveable
-
Save this model to the given path.
- save(String) - 类 中的方法org.apache.spark.sql.DataFrameWriter
-
Saves the content of the DataFrame at the specified path.
- save() - 类 中的方法org.apache.spark.sql.DataFrameWriter
-
Saves the content of the DataFrame as the specified table.
- Saveable - org.apache.spark.mllib.util中的接口
-
:: DeveloperApi ::
Trait for models and transformers which may be saved as files.
- saveAsHadoopDataset(JobConf) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Output the RDD to any Hadoop-supported storage system, using a Hadoop JobConf object for
that storage system.
- saveAsHadoopDataset(JobConf) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Output the RDD to any Hadoop-supported storage system, using a Hadoop JobConf object for
that storage system.
- saveAsHadoopFile(String, Class<?>, Class<?>, Class<F>, JobConf) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Output the RDD to any Hadoop-supported file system.
- saveAsHadoopFile(String, Class<?>, Class<?>, Class<F>) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Output the RDD to any Hadoop-supported file system.
- saveAsHadoopFile(String, Class<?>, Class<?>, Class<F>, Class<? extends CompressionCodec>) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Output the RDD to any Hadoop-supported file system, compressing with the supplied codec.
- saveAsHadoopFile(String, ClassTag<F>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Output the RDD to any Hadoop-supported file system, using a Hadoop OutputFormat class
supporting the key and value types K and V in this RDD.
- saveAsHadoopFile(String, Class<? extends CompressionCodec>, ClassTag<F>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Output the RDD to any Hadoop-supported file system, using a Hadoop OutputFormat class
supporting the key and value types K and V in this RDD.
- saveAsHadoopFile(String, Class<?>, Class<?>, Class<? extends OutputFormat<?, ?>>, Class<? extends CompressionCodec>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Output the RDD to any Hadoop-supported file system, using a Hadoop OutputFormat class
supporting the key and value types K and V in this RDD.
- saveAsHadoopFile(String, Class<?>, Class<?>, Class<? extends OutputFormat<?, ?>>, JobConf, Option<Class<? extends CompressionCodec>>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Output the RDD to any Hadoop-supported file system, using a Hadoop OutputFormat class
supporting the key and value types K and V in this RDD.
- saveAsHadoopFiles(String, String) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Save each RDD in this DStream as a Hadoop file.
- saveAsHadoopFiles(String, String, Class<?>, Class<?>, Class<F>) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Save each RDD in this DStream as a Hadoop file.
- saveAsHadoopFiles(String, String, Class<?>, Class<?>, Class<F>, JobConf) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Save each RDD in this DStream as a Hadoop file.
- saveAsHadoopFiles(String, String) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- saveAsHadoopFiles(String, String, Class<?>, Class<?>, Class<F>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- saveAsHadoopFiles(String, String, Class<?>, Class<?>, Class<F>, JobConf) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- saveAsHadoopFiles(String, String) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- saveAsHadoopFiles(String, String, Class<?>, Class<?>, Class<F>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- saveAsHadoopFiles(String, String, Class<?>, Class<?>, Class<F>, JobConf) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- saveAsHadoopFiles(String, String, ClassTag<F>) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Save each RDD in this DStream as a Hadoop file.
- saveAsHadoopFiles(String, String, Class<?>, Class<?>, Class<? extends OutputFormat<?, ?>>, JobConf) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Save each RDD in this DStream as a Hadoop file.
- saveAsLibSVMFile(RDD<LabeledPoint>, String) - 类 中的静态方法org.apache.spark.mllib.util.MLUtils
-
Save labeled data in LIBSVM format.
- saveAsNewAPIHadoopDataset(Configuration) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Output the RDD to any Hadoop-supported storage system, using
a Configuration object for that storage system.
- saveAsNewAPIHadoopDataset(Configuration) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Output the RDD to any Hadoop-supported storage system with new Hadoop API, using a Hadoop
Configuration object for that storage system.
- saveAsNewAPIHadoopFile(String, Class<?>, Class<?>, Class<F>, Configuration) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Output the RDD to any Hadoop-supported file system.
- saveAsNewAPIHadoopFile(String, Class<?>, Class<?>, Class<F>) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Output the RDD to any Hadoop-supported file system.
- saveAsNewAPIHadoopFile(String, ClassTag<F>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Output the RDD to any Hadoop-supported file system, using a new Hadoop API OutputFormat
(mapreduce.OutputFormat) object supporting the key and value types K and V in this RDD.
- saveAsNewAPIHadoopFile(String, Class<?>, Class<?>, Class<? extends OutputFormat<?, ?>>, Configuration) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Output the RDD to any Hadoop-supported file system, using a new Hadoop API OutputFormat
(mapreduce.OutputFormat) object supporting the key and value types K and V in this RDD.
- saveAsNewAPIHadoopFiles(String, String) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Save each RDD in this DStream as a Hadoop file.
- saveAsNewAPIHadoopFiles(String, String, Class<?>, Class<?>, Class<F>) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Save each RDD in this DStream as a Hadoop file.
- saveAsNewAPIHadoopFiles(String, String, Class<?>, Class<?>, Class<F>, Configuration) - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Save each RDD in this DStream as a Hadoop file.
- saveAsNewAPIHadoopFiles(String, String) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- saveAsNewAPIHadoopFiles(String, String, Class<?>, Class<?>, Class<F>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- saveAsNewAPIHadoopFiles(String, String, Class<?>, Class<?>, Class<F>, Configuration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- saveAsNewAPIHadoopFiles(String, String) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- saveAsNewAPIHadoopFiles(String, String, Class<?>, Class<?>, Class<F>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- saveAsNewAPIHadoopFiles(String, String, Class<?>, Class<?>, Class<F>, Configuration) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- saveAsNewAPIHadoopFiles(String, String, ClassTag<F>) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Save each RDD in this DStream as a Hadoop file.
- saveAsNewAPIHadoopFiles(String, String, Class<?>, Class<?>, Class<? extends OutputFormat<?, ?>>, Configuration) - 类 中的方法org.apache.spark.streaming.dstream.PairDStreamFunctions
-
Save each RDD in this DStream as a Hadoop file.
- saveAsNewAPIHadoopFiles$default$6() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- saveAsNewAPIHadoopFiles$default$6() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- saveAsObjectFile(String) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- saveAsObjectFile(String) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- saveAsObjectFile(String) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- saveAsObjectFile(String) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Save this RDD as a SequenceFile of serialized objects.
- saveAsObjectFile(String) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- saveAsObjectFile(String) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- saveAsObjectFile(String) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- saveAsObjectFile(String) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- saveAsObjectFile(String) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- saveAsObjectFile(String) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- saveAsObjectFile(String) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- saveAsObjectFile(String) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- saveAsObjectFile(String) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- saveAsObjectFile(String) - 类 中的方法org.apache.spark.rdd.RDD
-
Save this RDD as a SequenceFile of serialized objects.
- saveAsObjectFile(String) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- saveAsObjectFiles(String, String) - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Save each RDD in this DStream as a Sequence file of serialized objects.
- saveAsSequenceFile(String, Option<Class<? extends CompressionCodec>>) - 类 中的方法org.apache.spark.rdd.SequenceFileRDDFunctions
-
Output the RDD as a Hadoop SequenceFile using the Writable types we infer from the RDD's key
and value types.
- saveAsTable(String) - 类 中的方法org.apache.spark.sql.DataFrameWriter
-
Saves the content of the DataFrame as the specified table.
- saveAsTextFile(String) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- saveAsTextFile(String, Class<? extends CompressionCodec>) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- saveAsTextFile(String) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- saveAsTextFile(String, Class<? extends CompressionCodec>) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- saveAsTextFile(String) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- saveAsTextFile(String, Class<? extends CompressionCodec>) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- saveAsTextFile(String) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Save this RDD as a text file, using string representations of elements.
- saveAsTextFile(String, Class<? extends CompressionCodec>) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Save this RDD as a compressed text file, using string representations of elements.
- saveAsTextFile(String) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- saveAsTextFile(String, Class<? extends CompressionCodec>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- saveAsTextFile(String) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- saveAsTextFile(String, Class<? extends CompressionCodec>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- saveAsTextFile(String) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- saveAsTextFile(String, Class<? extends CompressionCodec>) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- saveAsTextFile(String) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- saveAsTextFile(String, Class<? extends CompressionCodec>) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- saveAsTextFile(String) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- saveAsTextFile(String, Class<? extends CompressionCodec>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- saveAsTextFile(String) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- saveAsTextFile(String, Class<? extends CompressionCodec>) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- saveAsTextFile(String) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- saveAsTextFile(String, Class<? extends CompressionCodec>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- saveAsTextFile(String) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- saveAsTextFile(String, Class<? extends CompressionCodec>) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- saveAsTextFile(String) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- saveAsTextFile(String, Class<? extends CompressionCodec>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- saveAsTextFile(String) - 类 中的方法org.apache.spark.rdd.RDD
-
Save this RDD as a text file, using string representations of elements.
- saveAsTextFile(String, Class<? extends CompressionCodec>) - 类 中的方法org.apache.spark.rdd.RDD
-
Save this RDD as a compressed text file, using string representations of elements.
- saveAsTextFile(String) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- saveAsTextFile(String, Class<? extends CompressionCodec>) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- saveAsTextFiles(String, String) - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Save each RDD in this DStream as at text file, using string representation
of elements.
- savedTasks() - 类 中的方法org.apache.spark.status.LiveStage
-
- saveImpl(Params, PipelineStage[], SparkContext, String) - 类 中的方法org.apache.spark.ml.Pipeline.SharedReadWrite$
-
Save metadata and stages for a
Pipeline or
PipelineModel
- save metadata to path/metadata
- save stages to stages/IDX_UID
- saveImpl(M, String, SparkSession, JsonAST.JObject) - 类 中的静态方法org.apache.spark.ml.tree.EnsembleModelReadWrite
-
Helper method for saving a tree ensemble to disk.
- SaveLoadV1_0$() - 类 的构造器org.apache.spark.mllib.classification.impl.GLMClassificationModel.SaveLoadV1_0$
-
- SaveLoadV1_0$() - 类 的构造器org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV1_0$
-
- SaveLoadV1_0$() - 类 的构造器org.apache.spark.mllib.clustering.BisectingKMeansModel.SaveLoadV1_0$
-
- SaveLoadV1_0$() - 类 的构造器org.apache.spark.mllib.clustering.KMeansModel.SaveLoadV1_0$
-
- SaveLoadV1_0$() - 类 的构造器org.apache.spark.mllib.clustering.PowerIterationClusteringModel.SaveLoadV1_0$
-
- SaveLoadV1_0$() - 类 的构造器org.apache.spark.mllib.feature.ChiSqSelectorModel.SaveLoadV1_0$
-
- SaveLoadV1_0$() - 类 的构造器org.apache.spark.mllib.fpm.FPGrowthModel.SaveLoadV1_0$
-
- SaveLoadV1_0$() - 类 的构造器org.apache.spark.mllib.fpm.PrefixSpanModel.SaveLoadV1_0$
-
- SaveLoadV1_0$() - 类 的构造器org.apache.spark.mllib.recommendation.MatrixFactorizationModel.SaveLoadV1_0$
-
- SaveLoadV1_0$() - 类 的构造器org.apache.spark.mllib.regression.impl.GLMRegressionModel.SaveLoadV1_0$
-
- SaveLoadV1_0$() - 类 的构造器org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$
-
- SaveLoadV2_0$() - 类 的构造器org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV2_0$
-
- SaveMode - org.apache.spark.sql中的枚举
-
SaveMode is used to specify the expected behavior of saving a DataFrame to a data source.
- sc() - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
- sc() - 类 中的方法org.apache.spark.sql.SQLImplicits.StringToColumn
-
- scal(double, Vector) - 类 中的静态方法org.apache.spark.ml.linalg.BLAS
-
x = a * x
- scal(double, Vector) - 类 中的静态方法org.apache.spark.mllib.linalg.BLAS
-
x = a * x
- scalaBoolean() - 类 中的静态方法org.apache.spark.sql.Encoders
-
An encoder for Scala's primitive boolean type.
- scalaByte() - 类 中的静态方法org.apache.spark.sql.Encoders
-
An encoder for Scala's primitive byte type.
- scalaDouble() - 类 中的静态方法org.apache.spark.sql.Encoders
-
An encoder for Scala's primitive double type.
- scalaFloat() - 类 中的静态方法org.apache.spark.sql.Encoders
-
An encoder for Scala's primitive float type.
- scalaInt() - 类 中的静态方法org.apache.spark.sql.Encoders
-
An encoder for Scala's primitive int type.
- scalaIntToJavaLong(DStream<Object>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- scalaIntToJavaLong(DStream<Object>) - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
- scalaIntToJavaLong(DStream<Object>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- scalaIntToJavaLong(DStream<Object>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- scalaIntToJavaLong(DStream<Object>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- scalaIntToJavaLong(DStream<Object>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- scalaIntToJavaLong(DStream<Object>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- scalaLong() - 类 中的静态方法org.apache.spark.sql.Encoders
-
An encoder for Scala's primitive long type.
- scalaShort() - 类 中的静态方法org.apache.spark.sql.Encoders
-
An encoder for Scala's primitive short type.
- scalaToJavaLong(JavaPairDStream<K, Object>, ClassTag<K>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- scalaVersion() - 类 中的方法org.apache.spark.status.api.v1.RuntimeInfo
-
- scale() - 类 中的方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- scale() - 类 中的方法org.apache.spark.ml.regression.LinearRegressionModel
-
- scale() - 类 中的方法org.apache.spark.mllib.random.GammaGenerator
-
- scale() - 类 中的方法org.apache.spark.sql.types.Decimal
-
- scale() - 类 中的方法org.apache.spark.sql.types.DecimalType
-
- scalingVec() - 类 中的方法org.apache.spark.ml.feature.ElementwiseProduct
-
the vector to multiply with input vectors
- scalingVec() - 类 中的方法org.apache.spark.mllib.feature.ElementwiseProduct
-
- scan(B, Function2<B, B, B>, CanBuildFrom<Repr, B, That>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- scanLeft(B, Function2<B, A, B>, CanBuildFrom<Repr, B, That>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- scanRight(B, Function2<A, B, B>, CanBuildFrom<Repr, B, That>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- SCHEDULED() - 类 中的静态方法org.apache.spark.streaming.scheduler.ReceiverState
-
- SCHEDULER_DELAY() - 类 中的静态方法org.apache.spark.status.TaskIndexNames
-
- SCHEDULER_DELAY() - 类 中的静态方法org.apache.spark.ui.jobs.TaskDetailsClassNames
-
- SCHEDULER_DELAY() - 类 中的静态方法org.apache.spark.ui.ToolTips
-
- SchedulerBackendUtils - org.apache.spark.scheduler.cluster中的类
-
- SchedulerBackendUtils() - 类 的构造器org.apache.spark.scheduler.cluster.SchedulerBackendUtils
-
- schedulerDelay() - 类 中的方法org.apache.spark.status.api.v1.TaskMetricDistributions
-
- schedulerDelay(TaskData) - 类 中的静态方法org.apache.spark.status.AppStatusUtils
-
- schedulerDelay(long, long, long, long, long, long) - 类 中的静态方法org.apache.spark.status.AppStatusUtils
-
- SchedulerPool - org.apache.spark.status中的类
-
- SchedulerPool(String) - 类 的构造器org.apache.spark.status.SchedulerPool
-
- schedulingDelay() - 类 中的方法org.apache.spark.status.api.v1.streaming.BatchInfo
-
- schedulingDelay() - 类 中的方法org.apache.spark.streaming.scheduler.BatchInfo
-
Time taken for the first job of this batch to start processing from the time this batch
was submitted to the streaming scheduler.
- SchedulingMode - org.apache.spark.scheduler中的类
-
"FAIR" and "FIFO" determines which policy is used
to order tasks amongst a Schedulable's sub-queues
"NONE" is used when the a Schedulable has no sub-queues.
- SchedulingMode() - 类 的构造器org.apache.spark.scheduler.SchedulingMode
-
- schedulingPool() - 类 中的方法org.apache.spark.status.api.v1.StageData
-
- schedulingPool() - 类 中的方法org.apache.spark.status.LiveStage
-
- schema(StructType) - 类 中的方法org.apache.spark.sql.DataFrameReader
-
Specifies the input schema.
- schema(String) - 类 中的方法org.apache.spark.sql.DataFrameReader
-
Specifies the schema by using the input DDL-formatted string.
- schema() - 类 中的方法org.apache.spark.sql.Dataset
-
Returns the schema of this Dataset.
- schema() - 接口 中的方法org.apache.spark.sql.Encoder
-
Returns the schema of encoding this type of object as a Row.
- schema() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- schema() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- schema() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- schema() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- schema() - 接口 中的方法org.apache.spark.sql.Row
-
Schema for the row.
- schema() - 类 中的方法org.apache.spark.sql.sources.BaseRelation
-
- schema(StructType) - 类 中的方法org.apache.spark.sql.streaming.DataStreamReader
-
Specifies the input schema.
- schema(String) - 类 中的方法org.apache.spark.sql.streaming.DataStreamReader
-
Specifies the schema by using the input DDL-formatted string.
- schemaLess() - 类 中的方法org.apache.spark.sql.hive.execution.HiveScriptIOSchema
-
- SchemaRelationProvider - org.apache.spark.sql.sources中的接口
-
Implemented by objects that produce relations for a specific kind of data source
with a given schema.
- schemaString() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- schemaString() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- schemaString() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- schemaString() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- SchemaUtils - org.apache.spark.ml.util中的类
-
Utils for handling schemas.
- SchemaUtils() - 类 的构造器org.apache.spark.ml.util.SchemaUtils
-
- SchemaUtils - org.apache.spark.sql.util中的类
-
Utils for handling schemas.
- SchemaUtils() - 类 的构造器org.apache.spark.sql.util.SchemaUtils
-
- scope() - 类 中的方法org.apache.spark.storage.RDDInfo
-
- scoreAndLabels() - 类 中的方法org.apache.spark.mllib.evaluation.BinaryClassificationMetrics
-
- scratch() - 类 中的方法org.apache.spark.mllib.optimization.NNLS.Workspace
-
- script() - 类 中的方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- ScriptTransformationExec - org.apache.spark.sql.hive.execution中的类
-
Transforms the input by forking and running the specified script.
- ScriptTransformationExec(Seq<Expression>, String, Seq<Attribute>, SparkPlan, HiveScriptIOSchema) - 类 的构造器org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- ScriptTransformationWriterThread - org.apache.spark.sql.hive.execution中的类
-
- ScriptTransformationWriterThread(Iterator<InternalRow>, Seq<DataType>, org.apache.spark.sql.catalyst.expressions.Projection, AbstractSerDe, ObjectInspector, HiveScriptIOSchema, OutputStream, Process, org.apache.spark.util.CircularBuffer, TaskContext, Configuration) - 类 的构造器org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread
-
- second(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Extracts the seconds as an integer from a given date/timestamp/string.
- seconds() - 类 中的静态方法org.apache.spark.scheduler.StatsReportListener
-
- seconds(long) - 类 中的静态方法org.apache.spark.streaming.Durations
-
- Seconds - org.apache.spark.streaming中的类
-
Helper object that creates instance of
Duration representing
a given number of seconds.
- Seconds() - 类 的构造器org.apache.spark.streaming.Seconds
-
- securityManager() - 类 中的方法org.apache.spark.SparkEnv
-
- seed() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- seed() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- seed() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- seed() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- seed() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- seed() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- seed() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- seed() - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeans
-
- seed() - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
- seed() - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- seed() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixture
-
- seed() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- seed() - 类 中的静态方法org.apache.spark.ml.clustering.KMeans
-
- seed() - 类 中的静态方法org.apache.spark.ml.clustering.KMeansModel
-
- seed() - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- seed() - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- seed() - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- seed() - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSH
-
- seed() - 类 中的静态方法org.apache.spark.ml.feature.Word2Vec
-
- seed() - 类 中的静态方法org.apache.spark.ml.feature.Word2VecModel
-
- seed() - 接口 中的方法org.apache.spark.ml.param.shared.HasSeed
-
Param for random seed.
- seed() - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- seed() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- seed() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- seed() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- seed() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- seed() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- seed() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- seed() - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidator
-
- seed() - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidatorModel
-
- seed() - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplit
-
- seed() - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplitModel
-
- seedParam() - 类 中的静态方法org.apache.spark.ml.image.SamplePathFilter
-
- segmentLength(Function1<A, Object>, int) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- select(Column...) - 类 中的方法org.apache.spark.sql.Dataset
-
Selects a set of column based expressions.
- select(String, String...) - 类 中的方法org.apache.spark.sql.Dataset
-
Selects a set of columns.
- select(Seq<Column>) - 类 中的方法org.apache.spark.sql.Dataset
-
Selects a set of column based expressions.
- select(String, Seq<String>) - 类 中的方法org.apache.spark.sql.Dataset
-
Selects a set of columns.
- select(TypedColumn<T, U1>) - 类 中的方法org.apache.spark.sql.Dataset
-
:: Experimental ::
Returns a new Dataset by computing the given
Column expression for each element.
- select(TypedColumn<T, U1>, TypedColumn<T, U2>) - 类 中的方法org.apache.spark.sql.Dataset
-
:: Experimental ::
Returns a new Dataset by computing the given
Column expressions for each element.
- select(TypedColumn<T, U1>, TypedColumn<T, U2>, TypedColumn<T, U3>) - 类 中的方法org.apache.spark.sql.Dataset
-
:: Experimental ::
Returns a new Dataset by computing the given
Column expressions for each element.
- select(TypedColumn<T, U1>, TypedColumn<T, U2>, TypedColumn<T, U3>, TypedColumn<T, U4>) - 类 中的方法org.apache.spark.sql.Dataset
-
:: Experimental ::
Returns a new Dataset by computing the given
Column expressions for each element.
- select(TypedColumn<T, U1>, TypedColumn<T, U2>, TypedColumn<T, U3>, TypedColumn<T, U4>, TypedColumn<T, U5>) - 类 中的方法org.apache.spark.sql.Dataset
-
:: Experimental ::
Returns a new Dataset by computing the given
Column expressions for each element.
- selectedFeatures() - 类 中的方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
list of indices to select (filter).
- selectedFeatures() - 类 中的方法org.apache.spark.mllib.feature.ChiSqSelectorModel
-
- selectExpr(String...) - 类 中的方法org.apache.spark.sql.Dataset
-
Selects a set of SQL expressions.
- selectExpr(Seq<String>) - 类 中的方法org.apache.spark.sql.Dataset
-
Selects a set of SQL expressions.
- selectorType() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelector
-
- selectorType() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- selectorType() - 类 中的方法org.apache.spark.mllib.feature.ChiSqSelector
-
- semanticHash() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- semanticHash() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- semanticHash() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- semanticHash() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- sender() - 类 中的方法org.apache.spark.storage.BlockManagerMessages.RegisterBlockManager
-
- sendToDst(A) - 类 中的方法org.apache.spark.graphx.EdgeContext
-
Sends a message to the destination vertex.
- sendToDst(A) - 类 中的方法org.apache.spark.graphx.impl.AggregatingEdgeContext
-
- sendToSrc(A) - 类 中的方法org.apache.spark.graphx.EdgeContext
-
Sends a message to the source vertex.
- sendToSrc(A) - 类 中的方法org.apache.spark.graphx.impl.AggregatingEdgeContext
-
- seq() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- seqToString(Seq<T>, Function1<T, String>) - 类 中的静态方法org.apache.spark.internal.config.ConfigHelpers
-
- sequence() - 类 中的方法org.apache.spark.mllib.fpm.PrefixSpan.FreqSequence
-
- sequenceFile(String, Class<K>, Class<V>, int) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
Get an RDD for a Hadoop SequenceFile with given key and value types.
- sequenceFile(String, Class<K>, Class<V>) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
Get an RDD for a Hadoop SequenceFile.
- sequenceFile(String, Class<K>, Class<V>, int) - 类 中的方法org.apache.spark.SparkContext
-
Get an RDD for a Hadoop SequenceFile with given key and value types.
- sequenceFile(String, Class<K>, Class<V>) - 类 中的方法org.apache.spark.SparkContext
-
Get an RDD for a Hadoop SequenceFile with given key and value types.
- sequenceFile(String, int, ClassTag<K>, ClassTag<V>, Function0<WritableConverter<K>>, Function0<WritableConverter<V>>) - 类 中的方法org.apache.spark.SparkContext
-
Version of sequenceFile() for types implicitly convertible to Writables through a
WritableConverter.
- SequenceFileRDDFunctions<K,V> - org.apache.spark.rdd中的类
-
Extra functions available on RDDs of (key, value) pairs to create a Hadoop SequenceFile,
through an implicit conversion.
- SequenceFileRDDFunctions(RDD<Tuple2<K, V>>, Class<? extends Writable>, Class<? extends Writable>, Function1<K, Writable>, ClassTag<K>, Function1<V, Writable>, ClassTag<V>) - 类 的构造器org.apache.spark.rdd.SequenceFileRDDFunctions
-
- SER_TIME() - 类 中的静态方法org.apache.spark.status.TaskIndexNames
-
- SerDe - org.apache.spark.api.r中的类
-
Utility functions to serialize, deserialize objects to / from R
- SerDe() - 类 的构造器org.apache.spark.api.r.SerDe
-
- SERDE() - 类 中的静态方法org.apache.spark.sql.hive.execution.HiveOptions
-
- serde() - 类 中的方法org.apache.spark.sql.hive.execution.HiveOptions
-
- serdeProperties() - 类 中的方法org.apache.spark.sql.hive.execution.HiveOptions
-
- SerializableMapWrapper(Map<A, B>) - 类 的构造器org.apache.spark.api.java.JavaUtils.SerializableMapWrapper
-
- SerializableWritable<T extends org.apache.hadoop.io.Writable> - org.apache.spark中的类
-
- SerializableWritable(T) - 类 的构造器org.apache.spark.SerializableWritable
-
- SerializationDebugger - org.apache.spark.serializer中的类
-
- SerializationDebugger() - 类 的构造器org.apache.spark.serializer.SerializationDebugger
-
- SerializationDebugger.ObjectStreamClassMethods - org.apache.spark.serializer中的类
-
An implicit class that allows us to call private methods of ObjectStreamClass.
- SerializationDebugger.ObjectStreamClassMethods$ - org.apache.spark.serializer中的类
-
- SerializationFormats - org.apache.spark.api.r中的类
-
- SerializationFormats() - 类 的构造器org.apache.spark.api.r.SerializationFormats
-
- SerializationStream - org.apache.spark.serializer中的类
-
:: DeveloperApi ::
A stream for writing serialized objects.
- SerializationStream() - 类 的构造器org.apache.spark.serializer.SerializationStream
-
- serialize(Vector) - 类 中的方法org.apache.spark.mllib.linalg.VectorUDT
-
- serialize(T, ClassTag<T>) - 类 中的方法org.apache.spark.serializer.DummySerializerInstance
-
- serialize(T, ClassTag<T>) - 类 中的方法org.apache.spark.serializer.SerializerInstance
-
- serialize(Object) - 类 中的方法org.apache.spark.util.kvstore.KVStoreSerializer
-
- serialize(T) - 类 中的静态方法org.apache.spark.util.Utils
-
Serialize an object using Java serialization
- SERIALIZED_R_DATA_SCHEMA() - 类 中的静态方法org.apache.spark.sql.api.r.SQLUtils
-
- serializedData() - 类 中的方法org.apache.spark.scheduler.local.StatusUpdate
-
- serializedMapStatus(org.apache.spark.broadcast.BroadcastManager, boolean, int) - 类 中的方法org.apache.spark.ShuffleStatus
-
Serializes the mapStatuses array into an efficient compressed format.
- SerializedMemoryEntry<T> - org.apache.spark.storage.memory中的类
-
- SerializedMemoryEntry(org.apache.spark.util.io.ChunkedByteBuffer, MemoryMode, ClassTag<T>) - 类 的构造器org.apache.spark.storage.memory.SerializedMemoryEntry
-
- Serializer - org.apache.spark.serializer中的类
-
:: DeveloperApi ::
A serializer.
- Serializer() - 类 的构造器org.apache.spark.serializer.Serializer
-
- serializer() - 类 中的方法org.apache.spark.ShuffleDependency
-
- serializer() - 类 中的方法org.apache.spark.SparkEnv
-
- SerializerInstance - org.apache.spark.serializer中的类
-
:: DeveloperApi ::
An instance of a serializer, for use by one thread at a time.
- SerializerInstance() - 类 的构造器org.apache.spark.serializer.SerializerInstance
-
- serializerManager() - 类 中的方法org.apache.spark.SparkEnv
-
- serializeStream(OutputStream) - 类 中的方法org.apache.spark.serializer.DummySerializerInstance
-
- serializeStream(OutputStream) - 类 中的方法org.apache.spark.serializer.SerializerInstance
-
- serializeViaNestedStream(OutputStream, SerializerInstance, Function1<SerializationStream, BoxedUnit>) - 类 中的静态方法org.apache.spark.util.Utils
-
Serialize via nested stream using specific serializer
- ServletParams(Function1<HttpServletRequest, T>, String, Function1<T, String>) - 类 的构造器org.apache.spark.ui.JettyUtils.ServletParams
-
- ServletParams$() - 类 的构造器org.apache.spark.ui.JettyUtils.ServletParams$
-
- session(SparkSession) - 类 中的静态方法org.apache.spark.ml.r.RWrappers
-
- session(SparkSession) - 类 中的方法org.apache.spark.ml.util.MLReader
-
- session(SparkSession) - 类 中的方法org.apache.spark.ml.util.MLWriter
-
- session() - 类 中的静态方法org.apache.spark.sql.hive.HiveSessionStateBuilder
-
- sessionCatalog() - 类 中的方法org.apache.spark.sql.hive.RelationConversions
-
- SessionConfigSupport - org.apache.spark.sql.sources.v2中的接口
-
- sessionState() - 类 中的方法org.apache.spark.sql.SparkSession
-
State isolated across sessions, including SQL configurations, temporary tables, registered
functions, and everything else that accepts a SQLConf.
- set(long, long, int, int, VD, VD, ED) - 类 中的方法org.apache.spark.graphx.impl.AggregatingEdgeContext
-
- Set() - 类 中的静态方法org.apache.spark.metrics.sink.StatsdMetricType
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.classification.OneVsRest
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.classification.OneVsRestModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeans
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixture
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.clustering.KMeans
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.clustering.KMeansModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.evaluation.ClusteringEvaluator
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.evaluation.RegressionEvaluator
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.Binarizer
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.Bucketizer
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelector
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.ColumnPruner
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizer
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizerModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.DCT
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.ElementwiseProduct
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.FeatureHasher
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.HashingTF
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.IDF
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.IDFModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.Imputer
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.ImputerModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.IndexToString
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.Interaction
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.MaxAbsScaler
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.MaxAbsScalerModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSH
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSHModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScaler
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScalerModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.NGram
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.Normalizer
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoder
-
已过时。
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderEstimator
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.PCA
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.PCAModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.PolynomialExpansion
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.QuantileDiscretizer
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.RegexTokenizer
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.RFormula
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.SQLTransformer
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.StandardScaler
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.StandardScalerModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.StopWordsRemover
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.StringIndexer
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.StringIndexerModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.Tokenizer
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.VectorAssembler
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.VectorAttributeRewriter
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexer
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexerModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.VectorSizeHint
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.VectorSlicer
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.Word2Vec
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.feature.Word2VecModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowth
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowthModel
-
- set(Param<T>, T) - 接口 中的方法org.apache.spark.ml.param.Params
-
Sets a parameter in the embedded param map.
- set(String, Object) - 接口 中的方法org.apache.spark.ml.param.Params
-
Sets a parameter (by name) in the embedded param map.
- set(ParamPair<?>) - 接口 中的方法org.apache.spark.ml.param.Params
-
Sets a parameter in the embedded param map.
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.Pipeline
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.PipelineModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.recommendation.ALSModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegression
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegressionModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidator
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidatorModel
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplit
-
- set(Param<T>, T) - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplitModel
-
- set(String, long, long) - 类 中的静态方法org.apache.spark.rdd.InputFileBlockHolder
-
Sets the thread-local input block.
- set(String, String) - 类 中的方法org.apache.spark.SparkConf
-
Set a configuration variable.
- set(SparkEnv) - 类 中的静态方法org.apache.spark.SparkEnv
-
- set(String, String) - 类 中的方法org.apache.spark.sql.RuntimeConfig
-
Sets the given Spark runtime configuration property.
- set(String, boolean) - 类 中的方法org.apache.spark.sql.RuntimeConfig
-
Sets the given Spark runtime configuration property.
- set(String, long) - 类 中的方法org.apache.spark.sql.RuntimeConfig
-
Sets the given Spark runtime configuration property.
- set(long) - 类 中的方法org.apache.spark.sql.types.Decimal
-
Set this Decimal to the given Long.
- set(int) - 类 中的方法org.apache.spark.sql.types.Decimal
-
Set this Decimal to the given Int.
- set(long, int, int) - 类 中的方法org.apache.spark.sql.types.Decimal
-
Set this Decimal to the given unscaled Long, with a given precision and scale.
- set(BigDecimal, int, int) - 类 中的方法org.apache.spark.sql.types.Decimal
-
Set this Decimal to the given BigDecimal value, with a given precision and scale.
- set(BigDecimal) - 类 中的方法org.apache.spark.sql.types.Decimal
-
Set this Decimal to the given BigDecimal value, inheriting its precision and scale.
- set(BigInteger) - 类 中的方法org.apache.spark.sql.types.Decimal
-
If the value is not in the range of long, convert it to BigDecimal and
the precision and scale are based on the converted value.
- set(Decimal) - 类 中的方法org.apache.spark.sql.types.Decimal
-
Set this Decimal to the given Decimal value.
- setAcceptsNull(boolean) - 类 中的静态方法org.apache.spark.serializer.JavaIterableWrapperSerializer
-
- setActive(SQLContext) - 类 中的静态方法org.apache.spark.sql.SQLContext
-
- setActiveSession(SparkSession) - 类 中的静态方法org.apache.spark.sql.SparkSession
-
Changes the SparkSession that will be returned in this thread and its children when
SparkSession.getOrCreate() is called.
- setAggregationDepth(int) - 类 中的方法org.apache.spark.ml.classification.LinearSVC
-
Suggested depth for treeAggregate (greater than or equal to 2).
- setAggregationDepth(int) - 类 中的方法org.apache.spark.ml.classification.LogisticRegression
-
Suggested depth for treeAggregate (greater than or equal to 2).
- setAggregationDepth(int) - 类 中的方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
Suggested depth for treeAggregate (greater than or equal to 2).
- setAggregationDepth(int) - 类 中的方法org.apache.spark.ml.regression.LinearRegression
-
Suggested depth for treeAggregate (greater than or equal to 2).
- setAggregator(Aggregator<K, V, C>) - 类 中的方法org.apache.spark.rdd.ShuffledRDD
-
Set aggregator for RDD's shuffle.
- setAlgo(String) - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
Sets Algorithm using a String.
- setAll(Traversable<Tuple2<String, String>>) - 类 中的方法org.apache.spark.SparkConf
-
Set multiple parameters together
- setAlpha(double) - 类 中的方法org.apache.spark.ml.recommendation.ALS
-
- setAlpha(Vector) - 类 中的方法org.apache.spark.mllib.clustering.LDA
-
Alias for setDocConcentration()
- setAlpha(double) - 类 中的方法org.apache.spark.mllib.clustering.LDA
-
Alias for setDocConcentration()
- setAlpha(double) - 类 中的方法org.apache.spark.mllib.recommendation.ALS
-
Sets the constant used in computing confidence in implicit ALS.
- setAppName(String) - 类 中的方法org.apache.spark.launcher.AbstractLauncher
-
Set the application name.
- setAppName(String) - 类 中的方法org.apache.spark.launcher.SparkLauncher
-
- setAppName(String) - 类 中的方法org.apache.spark.SparkConf
-
Set a name for your application.
- setAppResource(String) - 类 中的方法org.apache.spark.launcher.AbstractLauncher
-
Set the main application resource.
- setAppResource(String) - 类 中的方法org.apache.spark.launcher.SparkLauncher
-
- setBandwidth(double) - 类 中的方法org.apache.spark.mllib.stat.KernelDensity
-
Sets the bandwidth (standard deviation) of the Gaussian kernel (default: 1.0).
- setBeta(double) - 类 中的方法org.apache.spark.mllib.clustering.LDA
-
Alias for setTopicConcentration()
- setBinary(boolean) - 类 中的方法org.apache.spark.ml.feature.CountVectorizer
-
- setBinary(boolean) - 类 中的方法org.apache.spark.ml.feature.CountVectorizerModel
-
- setBinary(boolean) - 类 中的方法org.apache.spark.ml.feature.HashingTF
-
- setBinary(boolean) - 类 中的方法org.apache.spark.mllib.feature.HashingTF
-
If true, term frequency vector will be binary such that non-zero term counts will be set to 1
(default: false)
- setBlocks(int) - 类 中的方法org.apache.spark.mllib.recommendation.ALS
-
Set the number of blocks for both user blocks and product blocks to parallelize the computation
into; pass -1 for an auto-configured number of blocks.
- setBlockSize(int) - 类 中的方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
Sets the value of param blockSize.
- setBucketLength(double) - 类 中的方法org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- setCacheNodeIds(boolean) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- setCacheNodeIds(boolean) - 类 中的方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- setCacheNodeIds(boolean) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- setCacheNodeIds(boolean) - 类 中的方法org.apache.spark.ml.classification.GBTClassifier
-
- setCacheNodeIds(boolean) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- setCacheNodeIds(boolean) - 类 中的方法org.apache.spark.ml.classification.RandomForestClassifier
-
- setCacheNodeIds(boolean) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- setCacheNodeIds(boolean) - 类 中的方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- setCacheNodeIds(boolean) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- setCacheNodeIds(boolean) - 类 中的方法org.apache.spark.ml.regression.GBTRegressor
-
- setCacheNodeIds(boolean) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- setCacheNodeIds(boolean) - 类 中的方法org.apache.spark.ml.regression.RandomForestRegressor
-
- setCallSite(String) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
Pass-through to SparkContext.setCallSite.
- setCallSite(String) - 类 中的方法org.apache.spark.SparkContext
-
Set the thread-local property for overriding the call sites
of actions and RDDs.
- setCaseSensitive(boolean) - 类 中的方法org.apache.spark.ml.feature.StopWordsRemover
-
- setCategoricalCols(String[]) - 类 中的方法org.apache.spark.ml.feature.FeatureHasher
-
- setCategoricalFeaturesInfo(Map<Integer, Integer>) - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
Sets categoricalFeaturesInfo using a Java Map.
- setCensorCol(String) - 类 中的方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- setCheckpointDir(String) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
Set the directory under which RDDs are going to be checkpointed.
- setCheckpointDir(String) - 类 中的方法org.apache.spark.SparkContext
-
Set the directory under which RDDs are going to be checkpointed.
- setCheckpointInterval(int) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- setCheckpointInterval(int) - 类 中的方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
Specifies how often to checkpoint the cached node IDs.
- setCheckpointInterval(int) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- setCheckpointInterval(int) - 类 中的方法org.apache.spark.ml.classification.GBTClassifier
-
Specifies how often to checkpoint the cached node IDs.
- setCheckpointInterval(int) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- setCheckpointInterval(int) - 类 中的方法org.apache.spark.ml.classification.RandomForestClassifier
-
Specifies how often to checkpoint the cached node IDs.
- setCheckpointInterval(int) - 类 中的方法org.apache.spark.ml.clustering.LDA
-
- setCheckpointInterval(int) - 类 中的方法org.apache.spark.ml.recommendation.ALS
-
- setCheckpointInterval(int) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- setCheckpointInterval(int) - 类 中的方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
Specifies how often to checkpoint the cached node IDs.
- setCheckpointInterval(int) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- setCheckpointInterval(int) - 类 中的方法org.apache.spark.ml.regression.GBTRegressor
-
Specifies how often to checkpoint the cached node IDs.
- setCheckpointInterval(int) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- setCheckpointInterval(int) - 类 中的方法org.apache.spark.ml.regression.RandomForestRegressor
-
Specifies how often to checkpoint the cached node IDs.
- setCheckpointInterval(int) - 类 中的方法org.apache.spark.mllib.clustering.LDA
-
Parameter for set checkpoint interval (greater than or equal to 1) or disable checkpoint (-1).
- setCheckpointInterval(int) - 类 中的方法org.apache.spark.mllib.recommendation.ALS
-
:: DeveloperApi ::
Set period (in iterations) between checkpoints (default = 10).
- setCheckpointInterval(int) - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- setClassifier(Classifier<?, ?, ?>) - 类 中的方法org.apache.spark.ml.classification.OneVsRest
-
- setColdStartStrategy(String) - 类 中的方法org.apache.spark.ml.recommendation.ALS
-
- setColdStartStrategy(String) - 类 中的方法org.apache.spark.ml.recommendation.ALSModel
-
- setCollectSubModels(boolean) - 类 中的方法org.apache.spark.ml.tuning.CrossValidator
-
Whether to collect submodels when fitting.
- setCollectSubModels(boolean) - 类 中的方法org.apache.spark.ml.tuning.TrainValidationSplit
-
Whether to collect submodels when fitting.
- setConf(String, String) - 类 中的方法org.apache.spark.launcher.AbstractLauncher
-
Set a single configuration value for the application.
- setConf(String, String) - 类 中的方法org.apache.spark.launcher.SparkLauncher
-
- setConf(Configuration) - 类 中的方法org.apache.spark.ml.image.SamplePathFilter
-
- setConf(Properties) - 类 中的方法org.apache.spark.sql.SQLContext
-
Set Spark SQL configuration properties.
- setConf(String, String) - 类 中的方法org.apache.spark.sql.SQLContext
-
Set the given Spark SQL configuration property.
- setConfig(String, String) - 类 中的静态方法org.apache.spark.launcher.SparkLauncher
-
Set a configuration value for the launcher library.
- setConvergenceTol(double) - 类 中的方法org.apache.spark.mllib.clustering.GaussianMixture
-
Set the largest change in log-likelihood at which convergence is
considered to have occurred.
- setConvergenceTol(double) - 类 中的方法org.apache.spark.mllib.optimization.GradientDescent
-
Set the convergence tolerance.
- setConvergenceTol(double) - 类 中的方法org.apache.spark.mllib.optimization.LBFGS
-
Set the convergence tolerance of iterations for L-BFGS.
- setConvergenceTol(double) - 类 中的方法org.apache.spark.mllib.regression.StreamingLinearRegressionWithSGD
-
Set the convergence tolerance.
- setCurrentDatabase(String) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
Sets the current default database in this session.
- setCustomHostname(String) - 类 中的静态方法org.apache.spark.util.Utils
-
Allow setting a custom host name because when we run on Mesos we need to use the same
hostname it reports to the master.
- setDecayFactor(double) - 类 中的方法org.apache.spark.mllib.clustering.StreamingKMeans
-
Set the forgetfulness of the previous centroids.
- setDefault(Param<T>, T) - 接口 中的方法org.apache.spark.ml.param.Params
-
Sets a default value for a param.
- setDefault(Seq<ParamPair<?>>) - 接口 中的方法org.apache.spark.ml.param.Params
-
Sets default values for a list of params.
- setDefaultClassLoader(ClassLoader) - 类 中的静态方法org.apache.spark.serializer.KryoSerializer
-
- setDefaultClassLoader(ClassLoader) - 类 中的方法org.apache.spark.serializer.Serializer
-
Sets a class loader for the serializer to use in deserialization.
- setDefaultSession(SparkSession) - 类 中的静态方法org.apache.spark.sql.SparkSession
-
Sets the default SparkSession that is returned by the builder.
- setDegree(int) - 类 中的方法org.apache.spark.ml.feature.PolynomialExpansion
-
- setDeployMode(String) - 类 中的方法org.apache.spark.launcher.AbstractLauncher
-
Set the deploy mode for the application.
- setDeployMode(String) - 类 中的方法org.apache.spark.launcher.SparkLauncher
-
- setDocConcentration(double[]) - 类 中的方法org.apache.spark.ml.clustering.LDA
-
- setDocConcentration(double) - 类 中的方法org.apache.spark.ml.clustering.LDA
-
- setDocConcentration(Vector) - 类 中的方法org.apache.spark.mllib.clustering.LDA
-
Concentration parameter (commonly named "alpha") for the prior placed on documents'
distributions over topics ("theta").
- setDocConcentration(double) - 类 中的方法org.apache.spark.mllib.clustering.LDA
-
Replicates a Double docConcentration to create a symmetric prior.
- setDropLast(boolean) - 类 中的方法org.apache.spark.ml.feature.OneHotEncoder
-
已过时。
- setDropLast(boolean) - 类 中的方法org.apache.spark.ml.feature.OneHotEncoderEstimator
-
- setDropLast(boolean) - 类 中的方法org.apache.spark.ml.feature.OneHotEncoderModel
-
- setElasticNetParam(double) - 类 中的方法org.apache.spark.ml.classification.LogisticRegression
-
Set the ElasticNet mixing parameter.
- setElasticNetParam(double) - 类 中的方法org.apache.spark.ml.regression.LinearRegression
-
Set the ElasticNet mixing parameter.
- setEpsilon(double) - 类 中的方法org.apache.spark.ml.regression.LinearRegression
-
Sets the value of param epsilon.
- setEpsilon(double) - 类 中的方法org.apache.spark.mllib.clustering.KMeans
-
Set the distance threshold within which we've consider centers to have converged.
- setEstimator(Estimator<?>) - 类 中的方法org.apache.spark.ml.tuning.CrossValidator
-
- setEstimator(Estimator<?>) - 类 中的方法org.apache.spark.ml.tuning.TrainValidationSplit
-
- setEstimatorParamMaps(ParamMap[]) - 类 中的方法org.apache.spark.ml.tuning.CrossValidator
-
- setEstimatorParamMaps(ParamMap[]) - 类 中的方法org.apache.spark.ml.tuning.TrainValidationSplit
-
- setEvaluator(Evaluator) - 类 中的方法org.apache.spark.ml.tuning.CrossValidator
-
- setEvaluator(Evaluator) - 类 中的方法org.apache.spark.ml.tuning.TrainValidationSplit
-
- setExecutorEnv(String, String) - 类 中的方法org.apache.spark.SparkConf
-
Set an environment variable to be used when launching executors for this application.
- setExecutorEnv(Seq<Tuple2<String, String>>) - 类 中的方法org.apache.spark.SparkConf
-
Set multiple environment variables to be used when launching executors.
- setExecutorEnv(Tuple2<String, String>[]) - 类 中的方法org.apache.spark.SparkConf
-
Set multiple environment variables to be used when launching executors.
- setFamily(String) - 类 中的方法org.apache.spark.ml.classification.LogisticRegression
-
Sets the value of param family.
- setFamily(String) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
Sets the value of param family.
- setFdr(double) - 类 中的方法org.apache.spark.ml.feature.ChiSqSelector
-
- setFdr(double) - 类 中的方法org.apache.spark.mllib.feature.ChiSqSelector
-
- setFeatureIndex(int) - 类 中的方法org.apache.spark.ml.regression.IsotonicRegression
-
- setFeatureIndex(int) - 类 中的方法org.apache.spark.ml.regression.IsotonicRegressionModel
-
- setFeaturesCol(String) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- setFeaturesCol(String) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- setFeaturesCol(String) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- setFeaturesCol(String) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- setFeaturesCol(String) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- setFeaturesCol(String) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- setFeaturesCol(String) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- setFeaturesCol(String) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- setFeaturesCol(String) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- setFeaturesCol(String) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- setFeaturesCol(String) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- setFeaturesCol(String) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- setFeaturesCol(String) - 类 中的方法org.apache.spark.ml.classification.OneVsRest
-
- setFeaturesCol(String) - 类 中的方法org.apache.spark.ml.classification.OneVsRestModel
-
- setFeaturesCol(String) - 类 中的静态方法org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- setFeaturesCol(String) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- setFeaturesCol(String) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- setFeaturesCol(String) - 类 中的方法org.apache.spark.ml.clustering.BisectingKMeans
-
- setFeaturesCol(String) - 类 中的方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
- setFeaturesCol(String) - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- setFeaturesCol(String) - 类 中的方法org.apache.spark.ml.clustering.GaussianMixture
-
- setFeaturesCol(String) - 类 中的方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- setFeaturesCol(String) - 类 中的方法org.apache.spark.ml.clustering.KMeans
-
- setFeaturesCol(String) - 类 中的方法org.apache.spark.ml.clustering.KMeansModel
-
- setFeaturesCol(String) - 类 中的方法org.apache.spark.ml.clustering.LDA
-
The features for LDA should be a Vector representing the word counts in a document.
- setFeaturesCol(String) - 类 中的方法org.apache.spark.ml.clustering.LDAModel
-
The features for LDA should be a Vector representing the word counts in a document.
- setFeaturesCol(String) - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- setFeaturesCol(String) - 类 中的方法org.apache.spark.ml.evaluation.ClusteringEvaluator
-
- setFeaturesCol(String) - 类 中的方法org.apache.spark.ml.feature.ChiSqSelector
-
- setFeaturesCol(String) - 类 中的方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- setFeaturesCol(String) - 类 中的方法org.apache.spark.ml.feature.RFormula
-
- setFeaturesCol(String) - 类 中的方法org.apache.spark.ml.PredictionModel
-
- setFeaturesCol(String) - 类 中的方法org.apache.spark.ml.Predictor
-
- setFeaturesCol(String) - 类 中的方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- setFeaturesCol(String) - 类 中的方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- setFeaturesCol(String) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- setFeaturesCol(String) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- setFeaturesCol(String) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- setFeaturesCol(String) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- setFeaturesCol(String) - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- setFeaturesCol(String) - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- setFeaturesCol(String) - 类 中的方法org.apache.spark.ml.regression.IsotonicRegression
-
- setFeaturesCol(String) - 类 中的方法org.apache.spark.ml.regression.IsotonicRegressionModel
-
- setFeaturesCol(String) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- setFeaturesCol(String) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- setFeaturesCol(String) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- setFeaturesCol(String) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- setFeatureSubsetStrategy(String) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- setFeatureSubsetStrategy(String) - 类 中的方法org.apache.spark.ml.classification.GBTClassifier
-
- setFeatureSubsetStrategy(String) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- setFeatureSubsetStrategy(String) - 类 中的方法org.apache.spark.ml.classification.RandomForestClassifier
-
- setFeatureSubsetStrategy(String) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- setFeatureSubsetStrategy(String) - 类 中的方法org.apache.spark.ml.regression.GBTRegressor
-
- setFeatureSubsetStrategy(String) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- setFeatureSubsetStrategy(String) - 类 中的方法org.apache.spark.ml.regression.RandomForestRegressor
-
- setFinalRDDStorageLevel(StorageLevel) - 类 中的方法org.apache.spark.mllib.recommendation.ALS
-
:: DeveloperApi ::
Sets storage level for final RDDs (user/product used in MatrixFactorizationModel).
- setFinalStorageLevel(String) - 类 中的方法org.apache.spark.ml.recommendation.ALS
-
- setFitIntercept(boolean) - 类 中的方法org.apache.spark.ml.classification.LinearSVC
-
Whether to fit an intercept term.
- setFitIntercept(boolean) - 类 中的方法org.apache.spark.ml.classification.LogisticRegression
-
Whether to fit an intercept term.
- setFitIntercept(boolean) - 类 中的方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
Set if we should fit the intercept
Default is true.
- setFitIntercept(boolean) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
Sets if we should fit the intercept.
- setFitIntercept(boolean) - 类 中的方法org.apache.spark.ml.regression.LinearRegression
-
Set if we should fit the intercept.
- setForceIndexLabel(boolean) - 类 中的方法org.apache.spark.ml.feature.RFormula
-
- setFormula(String) - 类 中的方法org.apache.spark.ml.feature.RFormula
-
Sets the formula to use for this transformer.
- setFpr(double) - 类 中的方法org.apache.spark.ml.feature.ChiSqSelector
-
- setFpr(double) - 类 中的方法org.apache.spark.mllib.feature.ChiSqSelector
-
- setFwe(double) - 类 中的方法org.apache.spark.ml.feature.ChiSqSelector
-
- setFwe(double) - 类 中的方法org.apache.spark.mllib.feature.ChiSqSelector
-
- setGaps(boolean) - 类 中的方法org.apache.spark.ml.feature.RegexTokenizer
-
- setGenerics(Kryo, Class<?>[]) - 类 中的静态方法org.apache.spark.serializer.JavaIterableWrapperSerializer
-
- setGradient(Gradient) - 类 中的方法org.apache.spark.mllib.optimization.GradientDescent
-
Set the gradient function (of the loss function of one single data example)
to be used for SGD.
- setGradient(Gradient) - 类 中的方法org.apache.spark.mllib.optimization.LBFGS
-
Set the gradient function (of the loss function of one single data example)
to be used for L-BFGS.
- setHalfLife(double, String) - 类 中的方法org.apache.spark.mllib.clustering.StreamingKMeans
-
Set the half life and time unit ("batches" or "points").
- setHandleInvalid(String) - 类 中的方法org.apache.spark.ml.feature.Bucketizer
-
- setHandleInvalid(String) - 类 中的方法org.apache.spark.ml.feature.OneHotEncoderEstimator
-
- setHandleInvalid(String) - 类 中的方法org.apache.spark.ml.feature.OneHotEncoderModel
-
- setHandleInvalid(String) - 类 中的方法org.apache.spark.ml.feature.QuantileDiscretizer
-
- setHandleInvalid(String) - 类 中的方法org.apache.spark.ml.feature.RFormula
-
- setHandleInvalid(String) - 类 中的方法org.apache.spark.ml.feature.StringIndexer
-
- setHandleInvalid(String) - 类 中的方法org.apache.spark.ml.feature.StringIndexerModel
-
- setHandleInvalid(String) - 类 中的方法org.apache.spark.ml.feature.VectorIndexer
-
- setHandleInvalid(String) - 类 中的方法org.apache.spark.ml.feature.VectorSizeHint
-
- setHashAlgorithm(String) - 类 中的方法org.apache.spark.mllib.feature.HashingTF
-
Set the hash algorithm used when mapping term to integer.
- setIfMissing(String, String) - 类 中的方法org.apache.spark.SparkConf
-
Set a parameter if it isn't already configured
- setImmutable(boolean) - 类 中的静态方法org.apache.spark.serializer.JavaIterableWrapperSerializer
-
- setImplicitPrefs(boolean) - 类 中的方法org.apache.spark.ml.recommendation.ALS
-
- setImplicitPrefs(boolean) - 类 中的方法org.apache.spark.mllib.recommendation.ALS
-
Sets whether to use implicit preference.
- setImpurity(String) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- setImpurity(String) - 类 中的方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- setImpurity(String) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- setImpurity(String) - 类 中的方法org.apache.spark.ml.classification.GBTClassifier
-
The impurity setting is ignored for GBT models.
- setImpurity(String) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- setImpurity(String) - 类 中的方法org.apache.spark.ml.classification.RandomForestClassifier
-
- setImpurity(String) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- setImpurity(String) - 类 中的方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- setImpurity(String) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- setImpurity(String) - 类 中的方法org.apache.spark.ml.regression.GBTRegressor
-
The impurity setting is ignored for GBT models.
- setImpurity(String) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- setImpurity(String) - 类 中的方法org.apache.spark.ml.regression.RandomForestRegressor
-
- setImpurity(Impurity) - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- setIndices(int[]) - 类 中的方法org.apache.spark.ml.feature.VectorSlicer
-
- setInitialCenters(Vector[], double[]) - 类 中的方法org.apache.spark.mllib.clustering.StreamingKMeans
-
Specify initial centers directly.
- setInitializationMode(String) - 类 中的方法org.apache.spark.mllib.clustering.KMeans
-
Set the initialization algorithm.
- setInitializationMode(String) - 类 中的方法org.apache.spark.mllib.clustering.PowerIterationClustering
-
Set the initialization mode.
- setInitializationSteps(int) - 类 中的方法org.apache.spark.mllib.clustering.KMeans
-
Set the number of steps for the k-means|| initialization mode.
- setInitialModel(GaussianMixtureModel) - 类 中的方法org.apache.spark.mllib.clustering.GaussianMixture
-
Set the initial GMM starting point, bypassing the random initialization.
- setInitialModel(KMeansModel) - 类 中的方法org.apache.spark.mllib.clustering.KMeans
-
Set the initial starting point, bypassing the random initialization or k-means||
The condition model.k == this.k must be met, failure results
in an IllegalArgumentException.
- setInitialWeights(Vector) - 类 中的方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
Sets the value of param initialWeights.
- setInitialWeights(Vector) - 类 中的方法org.apache.spark.mllib.classification.StreamingLogisticRegressionWithSGD
-
Set the initial weights.
- setInitialWeights(Vector) - 类 中的方法org.apache.spark.mllib.regression.StreamingLinearRegressionWithSGD
-
Set the initial weights.
- setInitMode(String) - 类 中的方法org.apache.spark.ml.clustering.KMeans
-
- setInitSteps(int) - 类 中的方法org.apache.spark.ml.clustering.KMeans
-
- setInputCol(String) - 类 中的方法org.apache.spark.ml.feature.Binarizer
-
- setInputCol(String) - 类 中的方法org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- setInputCol(String) - 类 中的方法org.apache.spark.ml.feature.Bucketizer
-
- setInputCol(String) - 类 中的方法org.apache.spark.ml.feature.CountVectorizer
-
- setInputCol(String) - 类 中的方法org.apache.spark.ml.feature.CountVectorizerModel
-
- setInputCol(String) - 类 中的静态方法org.apache.spark.ml.feature.DCT
-
- setInputCol(String) - 类 中的静态方法org.apache.spark.ml.feature.ElementwiseProduct
-
- setInputCol(String) - 类 中的方法org.apache.spark.ml.feature.HashingTF
-
- setInputCol(String) - 类 中的方法org.apache.spark.ml.feature.IDF
-
- setInputCol(String) - 类 中的方法org.apache.spark.ml.feature.IDFModel
-
- setInputCol(String) - 类 中的方法org.apache.spark.ml.feature.IndexToString
-
- setInputCol(String) - 类 中的方法org.apache.spark.ml.feature.MaxAbsScaler
-
- setInputCol(String) - 类 中的方法org.apache.spark.ml.feature.MaxAbsScalerModel
-
- setInputCol(String) - 类 中的方法org.apache.spark.ml.feature.MinHashLSH
-
- setInputCol(String) - 类 中的方法org.apache.spark.ml.feature.MinMaxScaler
-
- setInputCol(String) - 类 中的方法org.apache.spark.ml.feature.MinMaxScalerModel
-
- setInputCol(String) - 类 中的静态方法org.apache.spark.ml.feature.NGram
-
- setInputCol(String) - 类 中的静态方法org.apache.spark.ml.feature.Normalizer
-
- setInputCol(String) - 类 中的方法org.apache.spark.ml.feature.OneHotEncoder
-
已过时。
- setInputCol(String) - 类 中的方法org.apache.spark.ml.feature.PCA
-
- setInputCol(String) - 类 中的方法org.apache.spark.ml.feature.PCAModel
-
- setInputCol(String) - 类 中的静态方法org.apache.spark.ml.feature.PolynomialExpansion
-
- setInputCol(String) - 类 中的方法org.apache.spark.ml.feature.QuantileDiscretizer
-
- setInputCol(String) - 类 中的静态方法org.apache.spark.ml.feature.RegexTokenizer
-
- setInputCol(String) - 类 中的方法org.apache.spark.ml.feature.StandardScaler
-
- setInputCol(String) - 类 中的方法org.apache.spark.ml.feature.StandardScalerModel
-
- setInputCol(String) - 类 中的方法org.apache.spark.ml.feature.StopWordsRemover
-
- setInputCol(String) - 类 中的方法org.apache.spark.ml.feature.StringIndexer
-
- setInputCol(String) - 类 中的方法org.apache.spark.ml.feature.StringIndexerModel
-
- setInputCol(String) - 类 中的静态方法org.apache.spark.ml.feature.Tokenizer
-
- setInputCol(String) - 类 中的方法org.apache.spark.ml.feature.VectorIndexer
-
- setInputCol(String) - 类 中的方法org.apache.spark.ml.feature.VectorIndexerModel
-
- setInputCol(String) - 类 中的方法org.apache.spark.ml.feature.VectorSizeHint
-
- setInputCol(String) - 类 中的方法org.apache.spark.ml.feature.VectorSlicer
-
- setInputCol(String) - 类 中的方法org.apache.spark.ml.feature.Word2Vec
-
- setInputCol(String) - 类 中的方法org.apache.spark.ml.feature.Word2VecModel
-
- setInputCol(String) - 类 中的方法org.apache.spark.ml.UnaryTransformer
-
- setInputCols(String[]) - 类 中的方法org.apache.spark.ml.feature.Bucketizer
-
- setInputCols(Seq<String>) - 类 中的方法org.apache.spark.ml.feature.FeatureHasher
-
- setInputCols(String[]) - 类 中的方法org.apache.spark.ml.feature.FeatureHasher
-
- setInputCols(String[]) - 类 中的方法org.apache.spark.ml.feature.Imputer
-
- setInputCols(String[]) - 类 中的方法org.apache.spark.ml.feature.ImputerModel
-
- setInputCols(String[]) - 类 中的方法org.apache.spark.ml.feature.Interaction
-
- setInputCols(String[]) - 类 中的方法org.apache.spark.ml.feature.OneHotEncoderEstimator
-
- setInputCols(String[]) - 类 中的方法org.apache.spark.ml.feature.OneHotEncoderModel
-
- setInputCols(String[]) - 类 中的方法org.apache.spark.ml.feature.QuantileDiscretizer
-
- setInputCols(String[]) - 类 中的方法org.apache.spark.ml.feature.VectorAssembler
-
- setIntercept(boolean) - 类 中的静态方法org.apache.spark.mllib.classification.LogisticRegressionWithSGD
-
- setIntercept(boolean) - 类 中的静态方法org.apache.spark.mllib.classification.SVMWithSGD
-
- setIntercept(boolean) - 类 中的方法org.apache.spark.mllib.regression.GeneralizedLinearAlgorithm
-
Set if the algorithm should add an intercept.
- setIntercept(boolean) - 类 中的静态方法org.apache.spark.mllib.regression.LassoWithSGD
-
- setIntercept(boolean) - 类 中的静态方法org.apache.spark.mllib.regression.LinearRegressionWithSGD
-
- setIntercept(boolean) - 类 中的静态方法org.apache.spark.mllib.regression.RidgeRegressionWithSGD
-
- setIntermediateRDDStorageLevel(StorageLevel) - 类 中的方法org.apache.spark.mllib.recommendation.ALS
-
:: DeveloperApi ::
Sets storage level for intermediate RDDs (user/product in/out links).
- setIntermediateStorageLevel(String) - 类 中的方法org.apache.spark.ml.recommendation.ALS
-
- setInverse(boolean) - 类 中的方法org.apache.spark.ml.feature.DCT
-
- setIsotonic(boolean) - 类 中的方法org.apache.spark.ml.regression.IsotonicRegression
-
- setIsotonic(boolean) - 类 中的方法org.apache.spark.mllib.regression.IsotonicRegression
-
Sets the isotonic parameter.
- setItemCol(String) - 类 中的方法org.apache.spark.ml.recommendation.ALS
-
- setItemCol(String) - 类 中的方法org.apache.spark.ml.recommendation.ALSModel
-
- setItemsCol(String) - 类 中的方法org.apache.spark.ml.fpm.FPGrowth
-
- setItemsCol(String) - 类 中的方法org.apache.spark.ml.fpm.FPGrowthModel
-
- setIterations(int) - 类 中的方法org.apache.spark.mllib.recommendation.ALS
-
Set the number of iterations to run.
- setJars(Seq<String>) - 类 中的方法org.apache.spark.SparkConf
-
Set JAR files to distribute to the cluster.
- setJars(String[]) - 类 中的方法org.apache.spark.SparkConf
-
Set JAR files to distribute to the cluster.
- setJavaHome(String) - 类 中的方法org.apache.spark.launcher.SparkLauncher
-
Set a custom JAVA_HOME for launching the Spark application.
- setJobDescription(String) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
Set a human readable description of the current job.
- setJobDescription(String) - 类 中的方法org.apache.spark.SparkContext
-
Set a human readable description of the current job.
- setJobGroup(String, String, boolean) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
Assigns a group ID to all the jobs started by this thread until the group ID is set to a
different value or cleared.
- setJobGroup(String, String) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
Assigns a group ID to all the jobs started by this thread until the group ID is set to a
different value or cleared.
- setJobGroup(String, String, boolean) - 类 中的方法org.apache.spark.SparkContext
-
Assigns a group ID to all the jobs started by this thread until the group ID is set to a
different value or cleared.
- setK(int) - 类 中的方法org.apache.spark.ml.clustering.BisectingKMeans
-
- setK(int) - 类 中的方法org.apache.spark.ml.clustering.GaussianMixture
-
- setK(int) - 类 中的方法org.apache.spark.ml.clustering.KMeans
-
- setK(int) - 类 中的方法org.apache.spark.ml.clustering.LDA
-
- setK(int) - 类 中的方法org.apache.spark.ml.feature.PCA
-
- setK(int) - 类 中的方法org.apache.spark.mllib.clustering.BisectingKMeans
-
Sets the desired number of leaf clusters (default: 4).
- setK(int) - 类 中的方法org.apache.spark.mllib.clustering.GaussianMixture
-
Set the number of Gaussians in the mixture model.
- setK(int) - 类 中的方法org.apache.spark.mllib.clustering.KMeans
-
Set the number of clusters to create (k).
- setK(int) - 类 中的方法org.apache.spark.mllib.clustering.LDA
-
Set the number of topics to infer, i.e., the number of soft cluster centers.
- setK(int) - 类 中的方法org.apache.spark.mllib.clustering.PowerIterationClustering
-
Set the number of clusters.
- setK(int) - 类 中的方法org.apache.spark.mllib.clustering.StreamingKMeans
-
Set the number of clusters.
- setKappa(double) - 类 中的方法org.apache.spark.mllib.clustering.OnlineLDAOptimizer
-
Learning rate: exponential decay rate---should be between
(0.5, 1.0] to guarantee asymptotic convergence.
- setKeepLastCheckpoint(boolean) - 类 中的方法org.apache.spark.ml.clustering.LDA
-
- setKeepLastCheckpoint(boolean) - 类 中的方法org.apache.spark.mllib.clustering.EMLDAOptimizer
-
If using checkpointing, this indicates whether to keep the last checkpoint (vs clean up).
- setKeyOrdering(Ordering<K>) - 类 中的方法org.apache.spark.rdd.ShuffledRDD
-
Set key ordering for RDD's shuffle.
- setLabelCol(String) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- setLabelCol(String) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- setLabelCol(String) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- setLabelCol(String) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- setLabelCol(String) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- setLabelCol(String) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- setLabelCol(String) - 类 中的方法org.apache.spark.ml.classification.OneVsRest
-
- setLabelCol(String) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- setLabelCol(String) - 类 中的方法org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
-
- setLabelCol(String) - 类 中的方法org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- setLabelCol(String) - 类 中的方法org.apache.spark.ml.evaluation.RegressionEvaluator
-
- setLabelCol(String) - 类 中的方法org.apache.spark.ml.feature.ChiSqSelector
-
- setLabelCol(String) - 类 中的方法org.apache.spark.ml.feature.RFormula
-
- setLabelCol(String) - 类 中的方法org.apache.spark.ml.Predictor
-
- setLabelCol(String) - 类 中的方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- setLabelCol(String) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- setLabelCol(String) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- setLabelCol(String) - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- setLabelCol(String) - 类 中的方法org.apache.spark.ml.regression.IsotonicRegression
-
- setLabelCol(String) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- setLabelCol(String) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- setLabels(String[]) - 类 中的方法org.apache.spark.ml.feature.IndexToString
-
- setLambda(double) - 类 中的方法org.apache.spark.mllib.classification.NaiveBayes
-
Set the smoothing parameter.
- setLambda(double) - 类 中的方法org.apache.spark.mllib.recommendation.ALS
-
Set the regularization parameter, lambda.
- setLayers(int[]) - 类 中的方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
Sets the value of param layers.
- setLearningDecay(double) - 类 中的方法org.apache.spark.ml.clustering.LDA
-
- setLearningOffset(double) - 类 中的方法org.apache.spark.ml.clustering.LDA
-
- setLearningRate(double) - 类 中的方法org.apache.spark.mllib.feature.Word2Vec
-
Sets initial learning rate (default: 0.025).
- setLearningRate(double) - 类 中的方法org.apache.spark.mllib.tree.configuration.BoostingStrategy
-
- setLink(String) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
Sets the value of param link.
- setLinkPower(double) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
Sets the value of param linkPower.
- setLinkPredictionCol(String) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
Sets the link prediction (linear predictor) column name.
- setLinkPredictionCol(String) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
Sets the link prediction (linear predictor) column name.
- setLocalProperty(String, String) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
Set a local property that affects jobs submitted from this thread, and all child
threads, such as the Spark fair scheduler pool.
- setLocalProperty(String, String) - 类 中的方法org.apache.spark.SparkContext
-
Set a local property that affects jobs submitted from this thread, such as the Spark fair
scheduler pool.
- setLogLevel(String) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
Control our logLevel.
- setLogLevel(String) - 类 中的方法org.apache.spark.SparkContext
-
Control our logLevel.
- setLogLevel(Level) - 类 中的静态方法org.apache.spark.util.Utils
-
configure a new log4j level
- setLoss(String) - 类 中的方法org.apache.spark.ml.regression.LinearRegression
-
Sets the value of param loss.
- setLoss(Loss) - 类 中的方法org.apache.spark.mllib.tree.configuration.BoostingStrategy
-
- setLossType(String) - 类 中的方法org.apache.spark.ml.classification.GBTClassifier
-
- setLossType(String) - 类 中的方法org.apache.spark.ml.regression.GBTRegressor
-
- setLowerBoundsOnCoefficients(Matrix) - 类 中的方法org.apache.spark.ml.classification.LogisticRegression
-
Set the lower bounds on coefficients if fitting under bound constrained optimization.
- setLowerBoundsOnIntercepts(Vector) - 类 中的方法org.apache.spark.ml.classification.LogisticRegression
-
Set the lower bounds on intercepts if fitting under bound constrained optimization.
- setMainClass(String) - 类 中的方法org.apache.spark.launcher.AbstractLauncher
-
Sets the application class name for Java/Scala applications.
- setMainClass(String) - 类 中的方法org.apache.spark.launcher.SparkLauncher
-
- setMapSideCombine(boolean) - 类 中的方法org.apache.spark.rdd.ShuffledRDD
-
Set mapSideCombine flag for RDD's shuffle.
- setMaster(String) - 类 中的方法org.apache.spark.launcher.AbstractLauncher
-
Set the Spark master for the application.
- setMaster(String) - 类 中的方法org.apache.spark.launcher.SparkLauncher
-
- setMaster(String) - 类 中的方法org.apache.spark.SparkConf
-
The master URL to connect to, such as "local" to run locally with one thread, "local[4]" to
run locally with 4 cores, or "spark://master:7077" to run on a Spark standalone cluster.
- setMax(double) - 类 中的方法org.apache.spark.ml.feature.MinMaxScaler
-
- setMax(double) - 类 中的方法org.apache.spark.ml.feature.MinMaxScalerModel
-
- setMaxBins(int) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- setMaxBins(int) - 类 中的方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- setMaxBins(int) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- setMaxBins(int) - 类 中的方法org.apache.spark.ml.classification.GBTClassifier
-
- setMaxBins(int) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- setMaxBins(int) - 类 中的方法org.apache.spark.ml.classification.RandomForestClassifier
-
- setMaxBins(int) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- setMaxBins(int) - 类 中的方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- setMaxBins(int) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- setMaxBins(int) - 类 中的方法org.apache.spark.ml.regression.GBTRegressor
-
- setMaxBins(int) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- setMaxBins(int) - 类 中的方法org.apache.spark.ml.regression.RandomForestRegressor
-
- setMaxBins(int) - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- setMaxCategories(int) - 类 中的方法org.apache.spark.ml.feature.VectorIndexer
-
- setMaxDepth(int) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- setMaxDepth(int) - 类 中的方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- setMaxDepth(int) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- setMaxDepth(int) - 类 中的方法org.apache.spark.ml.classification.GBTClassifier
-
- setMaxDepth(int) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- setMaxDepth(int) - 类 中的方法org.apache.spark.ml.classification.RandomForestClassifier
-
- setMaxDepth(int) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- setMaxDepth(int) - 类 中的方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- setMaxDepth(int) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- setMaxDepth(int) - 类 中的方法org.apache.spark.ml.regression.GBTRegressor
-
- setMaxDepth(int) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- setMaxDepth(int) - 类 中的方法org.apache.spark.ml.regression.RandomForestRegressor
-
- setMaxDepth(int) - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- setMaxIter(int) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- setMaxIter(int) - 类 中的方法org.apache.spark.ml.classification.GBTClassifier
-
- setMaxIter(int) - 类 中的方法org.apache.spark.ml.classification.LinearSVC
-
Set the maximum number of iterations.
- setMaxIter(int) - 类 中的方法org.apache.spark.ml.classification.LogisticRegression
-
Set the maximum number of iterations.
- setMaxIter(int) - 类 中的方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
Set the maximum number of iterations.
- setMaxIter(int) - 类 中的方法org.apache.spark.ml.clustering.BisectingKMeans
-
- setMaxIter(int) - 类 中的方法org.apache.spark.ml.clustering.GaussianMixture
-
- setMaxIter(int) - 类 中的方法org.apache.spark.ml.clustering.KMeans
-
- setMaxIter(int) - 类 中的方法org.apache.spark.ml.clustering.LDA
-
- setMaxIter(int) - 类 中的方法org.apache.spark.ml.feature.Word2Vec
-
- setMaxIter(int) - 类 中的方法org.apache.spark.ml.recommendation.ALS
-
- setMaxIter(int) - 类 中的方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
Set the maximum number of iterations.
- setMaxIter(int) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- setMaxIter(int) - 类 中的方法org.apache.spark.ml.regression.GBTRegressor
-
- setMaxIter(int) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
Sets the maximum number of iterations (applicable for solver "irls").
- setMaxIter(int) - 类 中的方法org.apache.spark.ml.regression.LinearRegression
-
Set the maximum number of iterations.
- setMaxIterations(int) - 类 中的方法org.apache.spark.mllib.clustering.BisectingKMeans
-
Sets the max number of k-means iterations to split clusters (default: 20).
- setMaxIterations(int) - 类 中的方法org.apache.spark.mllib.clustering.GaussianMixture
-
Set the maximum number of iterations allowed.
- setMaxIterations(int) - 类 中的方法org.apache.spark.mllib.clustering.KMeans
-
Set maximum number of iterations allowed.
- setMaxIterations(int) - 类 中的方法org.apache.spark.mllib.clustering.LDA
-
Set the maximum number of iterations allowed.
- setMaxIterations(int) - 类 中的方法org.apache.spark.mllib.clustering.PowerIterationClustering
-
Set maximum number of iterations of the power iteration loop
- setMaxLocalProjDBSize(long) - 类 中的方法org.apache.spark.mllib.fpm.PrefixSpan
-
Sets the maximum number of items (including delimiters used in the internal storage format)
allowed in a projected database before local processing (default: 32000000L).
- setMaxMemoryInMB(int) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- setMaxMemoryInMB(int) - 类 中的方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- setMaxMemoryInMB(int) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- setMaxMemoryInMB(int) - 类 中的方法org.apache.spark.ml.classification.GBTClassifier
-
- setMaxMemoryInMB(int) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- setMaxMemoryInMB(int) - 类 中的方法org.apache.spark.ml.classification.RandomForestClassifier
-
- setMaxMemoryInMB(int) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- setMaxMemoryInMB(int) - 类 中的方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- setMaxMemoryInMB(int) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- setMaxMemoryInMB(int) - 类 中的方法org.apache.spark.ml.regression.GBTRegressor
-
- setMaxMemoryInMB(int) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- setMaxMemoryInMB(int) - 类 中的方法org.apache.spark.ml.regression.RandomForestRegressor
-
- setMaxMemoryInMB(int) - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- setMaxPatternLength(int) - 类 中的方法org.apache.spark.mllib.fpm.PrefixSpan
-
Sets maximal pattern length (default: 10).
- setMaxSentenceLength(int) - 类 中的方法org.apache.spark.ml.feature.Word2Vec
-
- setMaxSentenceLength(int) - 类 中的方法org.apache.spark.mllib.feature.Word2Vec
-
Sets the maximum length (in words) of each sentence in the input data.
- setMetadata(Object) - 类 中的方法org.apache.spark.util.kvstore.InMemoryStore
-
- setMetadata(Object) - 接口 中的方法org.apache.spark.util.kvstore.KVStore
-
Writes the given value in the store metadata key.
- setMetadata(Object) - 类 中的方法org.apache.spark.util.kvstore.LevelDB
-
- setMetricName(String) - 类 中的方法org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
-
- setMetricName(String) - 类 中的方法org.apache.spark.ml.evaluation.ClusteringEvaluator
-
- setMetricName(String) - 类 中的方法org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- setMetricName(String) - 类 中的方法org.apache.spark.ml.evaluation.RegressionEvaluator
-
- setMin(double) - 类 中的方法org.apache.spark.ml.feature.MinMaxScaler
-
- setMin(double) - 类 中的方法org.apache.spark.ml.feature.MinMaxScalerModel
-
- setMinConfidence(double) - 类 中的方法org.apache.spark.ml.fpm.FPGrowth
-
- setMinConfidence(double) - 类 中的方法org.apache.spark.ml.fpm.FPGrowthModel
-
- setMinConfidence(double) - 类 中的方法org.apache.spark.mllib.fpm.AssociationRules
-
Sets the minimal confidence (default: 0.8).
- setMinCount(int) - 类 中的方法org.apache.spark.ml.feature.Word2Vec
-
- setMinCount(int) - 类 中的方法org.apache.spark.mllib.feature.Word2Vec
-
Sets minCount, the minimum number of times a token must appear to be included in the word2vec
model's vocabulary (default: 5).
- setMinDF(double) - 类 中的方法org.apache.spark.ml.feature.CountVectorizer
-
- setMinDivisibleClusterSize(double) - 类 中的方法org.apache.spark.ml.clustering.BisectingKMeans
-
- setMinDivisibleClusterSize(double) - 类 中的方法org.apache.spark.mllib.clustering.BisectingKMeans
-
Sets the minimum number of points (if greater than or equal to 1.0) or the minimum proportion
of points (if less than 1.0) of a divisible cluster (default: 1).
- setMinDocFreq(int) - 类 中的方法org.apache.spark.ml.feature.IDF
-
- setMiniBatchFraction(double) - 类 中的方法org.apache.spark.mllib.classification.StreamingLogisticRegressionWithSGD
-
Set the fraction of each batch to use for updates.
- setMiniBatchFraction(double) - 类 中的方法org.apache.spark.mllib.clustering.OnlineLDAOptimizer
-
Mini-batch fraction in (0, 1], which sets the fraction of document sampled and used in
each iteration.
- setMiniBatchFraction(double) - 类 中的方法org.apache.spark.mllib.optimization.GradientDescent
-
Set fraction of data to be used for each SGD iteration.
- setMiniBatchFraction(double) - 类 中的方法org.apache.spark.mllib.regression.StreamingLinearRegressionWithSGD
-
Set the fraction of each batch to use for updates.
- setMinInfoGain(double) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- setMinInfoGain(double) - 类 中的方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- setMinInfoGain(double) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- setMinInfoGain(double) - 类 中的方法org.apache.spark.ml.classification.GBTClassifier
-
- setMinInfoGain(double) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- setMinInfoGain(double) - 类 中的方法org.apache.spark.ml.classification.RandomForestClassifier
-
- setMinInfoGain(double) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- setMinInfoGain(double) - 类 中的方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- setMinInfoGain(double) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- setMinInfoGain(double) - 类 中的方法org.apache.spark.ml.regression.GBTRegressor
-
- setMinInfoGain(double) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- setMinInfoGain(double) - 类 中的方法org.apache.spark.ml.regression.RandomForestRegressor
-
- setMinInfoGain(double) - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- setMinInstancesPerNode(int) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- setMinInstancesPerNode(int) - 类 中的方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- setMinInstancesPerNode(int) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- setMinInstancesPerNode(int) - 类 中的方法org.apache.spark.ml.classification.GBTClassifier
-
- setMinInstancesPerNode(int) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- setMinInstancesPerNode(int) - 类 中的方法org.apache.spark.ml.classification.RandomForestClassifier
-
- setMinInstancesPerNode(int) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- setMinInstancesPerNode(int) - 类 中的方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- setMinInstancesPerNode(int) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- setMinInstancesPerNode(int) - 类 中的方法org.apache.spark.ml.regression.GBTRegressor
-
- setMinInstancesPerNode(int) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- setMinInstancesPerNode(int) - 类 中的方法org.apache.spark.ml.regression.RandomForestRegressor
-
- setMinInstancesPerNode(int) - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- setMinSupport(double) - 类 中的方法org.apache.spark.ml.fpm.FPGrowth
-
- setMinSupport(double) - 类 中的方法org.apache.spark.mllib.fpm.FPGrowth
-
Sets the minimal support level (default: 0.3).
- setMinSupport(double) - 类 中的方法org.apache.spark.mllib.fpm.PrefixSpan
-
Sets the minimal support level (default: 0.1).
- setMinTF(double) - 类 中的方法org.apache.spark.ml.feature.CountVectorizer
-
- setMinTF(double) - 类 中的方法org.apache.spark.ml.feature.CountVectorizerModel
-
- setMinTokenLength(int) - 类 中的方法org.apache.spark.ml.feature.RegexTokenizer
-
- setMissingValue(double) - 类 中的方法org.apache.spark.ml.feature.Imputer
-
- setModelType(String) - 类 中的方法org.apache.spark.ml.classification.NaiveBayes
-
Set the model type using a string (case-sensitive).
- setModelType(String) - 类 中的方法org.apache.spark.mllib.classification.NaiveBayes
-
Set the model type using a string (case-sensitive).
- setN(int) - 类 中的方法org.apache.spark.ml.feature.NGram
-
- setName(String) - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
Assign a name to this RDD
- setName(String) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Assign a name to this RDD
- setName(String) - 类 中的方法org.apache.spark.api.java.JavaRDD
-
Assign a name to this RDD
- setName(String) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- setName(String) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- setName(String) - 类 中的方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- setName(String) - 类 中的方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- setName(String) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- setName(String) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- setName(String) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- setName(String) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- setName(String) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- setName(String) - 类 中的方法org.apache.spark.rdd.RDD
-
Assign a name to this RDD
- setName(String) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- setNames(String[]) - 类 中的方法org.apache.spark.ml.feature.VectorSlicer
-
- setNonnegative(boolean) - 类 中的方法org.apache.spark.ml.recommendation.ALS
-
- setNonnegative(boolean) - 类 中的方法org.apache.spark.mllib.recommendation.ALS
-
Set whether the least-squares problems solved at each iteration should have
nonnegativity constraints.
- setNullAt(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarArray
-
- setNullAt(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarRow
-
- setNumBlocks(int) - 类 中的方法org.apache.spark.ml.recommendation.ALS
-
Sets both numUserBlocks and numItemBlocks to the specific value.
- setNumBuckets(int) - 类 中的方法org.apache.spark.ml.feature.QuantileDiscretizer
-
- setNumBucketsArray(int[]) - 类 中的方法org.apache.spark.ml.feature.QuantileDiscretizer
-
- setNumClasses(int) - 类 中的方法org.apache.spark.mllib.classification.LogisticRegressionWithLBFGS
-
Set the number of possible outcomes for k classes classification problem in
Multinomial Logistic Regression.
- setNumClasses(int) - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- setNumCorrections(int) - 类 中的方法org.apache.spark.mllib.optimization.LBFGS
-
Set the number of corrections used in the LBFGS update.
- setNumFeatures(int) - 类 中的方法org.apache.spark.ml.feature.FeatureHasher
-
- setNumFeatures(int) - 类 中的方法org.apache.spark.ml.feature.HashingTF
-
- setNumFolds(int) - 类 中的方法org.apache.spark.ml.tuning.CrossValidator
-
- setNumHashTables(int) - 类 中的方法org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- setNumHashTables(int) - 类 中的方法org.apache.spark.ml.feature.MinHashLSH
-
- setNumItemBlocks(int) - 类 中的方法org.apache.spark.ml.recommendation.ALS
-
- setNumIterations(int) - 类 中的方法org.apache.spark.mllib.classification.StreamingLogisticRegressionWithSGD
-
Set the number of iterations of gradient descent to run per update.
- setNumIterations(int) - 类 中的方法org.apache.spark.mllib.feature.Word2Vec
-
Sets number of iterations (default: 1), which should be smaller than or equal to number of
partitions.
- setNumIterations(int) - 类 中的方法org.apache.spark.mllib.optimization.GradientDescent
-
Set the number of iterations for SGD.
- setNumIterations(int) - 类 中的方法org.apache.spark.mllib.optimization.LBFGS
-
Set the maximal number of iterations for L-BFGS.
- setNumIterations(int) - 类 中的方法org.apache.spark.mllib.regression.StreamingLinearRegressionWithSGD
-
Set the number of iterations of gradient descent to run per update.
- setNumIterations(int) - 类 中的方法org.apache.spark.mllib.tree.configuration.BoostingStrategy
-
- setNumPartitions(int) - 类 中的方法org.apache.spark.ml.feature.Word2Vec
-
- setNumPartitions(int) - 类 中的方法org.apache.spark.ml.fpm.FPGrowth
-
- setNumPartitions(int) - 类 中的方法org.apache.spark.mllib.feature.Word2Vec
-
Sets number of partitions (default: 1).
- setNumPartitions(int) - 类 中的方法org.apache.spark.mllib.fpm.FPGrowth
-
Sets the number of partitions used by parallel FP-growth (default: same as input data).
- setNumRows(int) - 类 中的方法org.apache.spark.sql.vectorized.ColumnarBatch
-
Sets the number of rows in this batch.
- setNumTopFeatures(int) - 类 中的方法org.apache.spark.ml.feature.ChiSqSelector
-
- setNumTopFeatures(int) - 类 中的方法org.apache.spark.mllib.feature.ChiSqSelector
-
- setNumTrees(int) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- setNumTrees(int) - 类 中的方法org.apache.spark.ml.classification.RandomForestClassifier
-
- setNumTrees(int) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- setNumTrees(int) - 类 中的方法org.apache.spark.ml.regression.RandomForestRegressor
-
- setNumUserBlocks(int) - 类 中的方法org.apache.spark.ml.recommendation.ALS
-
- setOffsetCol(String) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
Sets the value of param offsetCol.
- setOffsetRange(Optional<Offset>, Optional<Offset>) - 接口 中的方法org.apache.spark.sql.sources.v2.reader.streaming.MicroBatchReader
-
Set the desired offset range for reader factories created from this reader.
- setOptimizeDocConcentration(boolean) - 类 中的方法org.apache.spark.ml.clustering.LDA
-
- setOptimizeDocConcentration(boolean) - 类 中的方法org.apache.spark.mllib.clustering.OnlineLDAOptimizer
-
Sets whether to optimize docConcentration parameter during training.
- setOptimizer(String) - 类 中的方法org.apache.spark.ml.clustering.LDA
-
- setOptimizer(LDAOptimizer) - 类 中的方法org.apache.spark.mllib.clustering.LDA
-
:: DeveloperApi ::
LDAOptimizer used to perform the actual calculation (default = EMLDAOptimizer)
- setOptimizer(String) - 类 中的方法org.apache.spark.mllib.clustering.LDA
-
Set the LDAOptimizer used to perform the actual calculation by algorithm name.
- setOrNull(long, int, int) - 类 中的方法org.apache.spark.sql.types.Decimal
-
Set this Decimal to the given unscaled Long, with a given precision and scale,
and return it, or return null if it cannot be set due to overflow.
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.Binarizer
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.Bucketizer
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.ChiSqSelector
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.CountVectorizer
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.CountVectorizerModel
-
- setOutputCol(String) - 类 中的静态方法org.apache.spark.ml.feature.DCT
-
- setOutputCol(String) - 类 中的静态方法org.apache.spark.ml.feature.ElementwiseProduct
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.FeatureHasher
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.HashingTF
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.IDF
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.IDFModel
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.IndexToString
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.Interaction
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.MaxAbsScaler
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.MaxAbsScalerModel
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.MinHashLSH
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.MinMaxScaler
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.MinMaxScalerModel
-
- setOutputCol(String) - 类 中的静态方法org.apache.spark.ml.feature.NGram
-
- setOutputCol(String) - 类 中的静态方法org.apache.spark.ml.feature.Normalizer
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.OneHotEncoder
-
已过时。
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.PCA
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.PCAModel
-
- setOutputCol(String) - 类 中的静态方法org.apache.spark.ml.feature.PolynomialExpansion
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.QuantileDiscretizer
-
- setOutputCol(String) - 类 中的静态方法org.apache.spark.ml.feature.RegexTokenizer
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.StandardScaler
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.StandardScalerModel
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.StopWordsRemover
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.StringIndexer
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.StringIndexerModel
-
- setOutputCol(String) - 类 中的静态方法org.apache.spark.ml.feature.Tokenizer
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.VectorAssembler
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.VectorIndexer
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.VectorIndexerModel
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.VectorSlicer
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.Word2Vec
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.feature.Word2VecModel
-
- setOutputCol(String) - 类 中的方法org.apache.spark.ml.UnaryTransformer
-
- setOutputCols(String[]) - 类 中的方法org.apache.spark.ml.feature.Bucketizer
-
- setOutputCols(String[]) - 类 中的方法org.apache.spark.ml.feature.Imputer
-
- setOutputCols(String[]) - 类 中的方法org.apache.spark.ml.feature.ImputerModel
-
- setOutputCols(String[]) - 类 中的方法org.apache.spark.ml.feature.OneHotEncoderEstimator
-
- setOutputCols(String[]) - 类 中的方法org.apache.spark.ml.feature.OneHotEncoderModel
-
- setOutputCols(String[]) - 类 中的方法org.apache.spark.ml.feature.QuantileDiscretizer
-
- setP(double) - 类 中的方法org.apache.spark.ml.feature.Normalizer
-
- setParallelism(int) - 类 中的方法org.apache.spark.ml.classification.OneVsRest
-
The implementation of parallel one vs. rest runs the classification for
each class in a separate threads.
- setParallelism(int) - 类 中的方法org.apache.spark.ml.tuning.CrossValidator
-
Set the maximum level of parallelism to evaluate models in parallel.
- setParallelism(int) - 类 中的方法org.apache.spark.ml.tuning.TrainValidationSplit
-
Set the mamixum level of parallelism to evaluate models in parallel.
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.classification.OneVsRestModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.clustering.KMeansModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.Bucketizer
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizerModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.IDFModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.ImputerModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.MaxAbsScalerModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSHModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScalerModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.PCAModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.StandardScalerModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.StringIndexerModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexerModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.feature.Word2VecModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowthModel
-
- setParent(Estimator<M>) - 类 中的方法org.apache.spark.ml.Model
-
Sets the parent of this model (Java API).
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.PipelineModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.recommendation.ALSModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegressionModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidatorModel
-
- setParent(Estimator<M>) - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplitModel
-
- setPattern(String) - 类 中的方法org.apache.spark.ml.feature.RegexTokenizer
-
- setPeacePeriod(int) - 类 中的方法org.apache.spark.mllib.stat.test.StreamingTest
-
Set the number of initial batches to ignore.
- setPercentile(double) - 类 中的方法org.apache.spark.ml.feature.ChiSqSelector
-
- setPercentile(double) - 类 中的方法org.apache.spark.mllib.feature.ChiSqSelector
-
- setPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- setPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- setPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- setPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- setPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- setPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- setPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- setPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- setPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- setPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- setPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- setPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- setPredictionCol(String) - 类 中的方法org.apache.spark.ml.classification.OneVsRest
-
- setPredictionCol(String) - 类 中的方法org.apache.spark.ml.classification.OneVsRestModel
-
- setPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- setPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- setPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- setPredictionCol(String) - 类 中的方法org.apache.spark.ml.clustering.BisectingKMeans
-
- setPredictionCol(String) - 类 中的方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
- setPredictionCol(String) - 类 中的方法org.apache.spark.ml.clustering.GaussianMixture
-
- setPredictionCol(String) - 类 中的方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- setPredictionCol(String) - 类 中的方法org.apache.spark.ml.clustering.KMeans
-
- setPredictionCol(String) - 类 中的方法org.apache.spark.ml.clustering.KMeansModel
-
- setPredictionCol(String) - 类 中的方法org.apache.spark.ml.evaluation.ClusteringEvaluator
-
- setPredictionCol(String) - 类 中的方法org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- setPredictionCol(String) - 类 中的方法org.apache.spark.ml.evaluation.RegressionEvaluator
-
- setPredictionCol(String) - 类 中的方法org.apache.spark.ml.fpm.FPGrowth
-
- setPredictionCol(String) - 类 中的方法org.apache.spark.ml.fpm.FPGrowthModel
-
- setPredictionCol(String) - 类 中的方法org.apache.spark.ml.PredictionModel
-
- setPredictionCol(String) - 类 中的方法org.apache.spark.ml.Predictor
-
- setPredictionCol(String) - 类 中的方法org.apache.spark.ml.recommendation.ALS
-
- setPredictionCol(String) - 类 中的方法org.apache.spark.ml.recommendation.ALSModel
-
- setPredictionCol(String) - 类 中的方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- setPredictionCol(String) - 类 中的方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- setPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- setPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- setPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- setPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- setPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- setPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- setPredictionCol(String) - 类 中的方法org.apache.spark.ml.regression.IsotonicRegression
-
- setPredictionCol(String) - 类 中的方法org.apache.spark.ml.regression.IsotonicRegressionModel
-
- setPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- setPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- setPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- setPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- setProbabilityCol(String) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- setProbabilityCol(String) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- setProbabilityCol(String) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- setProbabilityCol(String) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- setProbabilityCol(String) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- setProbabilityCol(String) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- setProbabilityCol(String) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- setProbabilityCol(String) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- setProbabilityCol(String) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- setProbabilityCol(String) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- setProbabilityCol(String) - 类 中的方法org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- setProbabilityCol(String) - 类 中的方法org.apache.spark.ml.classification.ProbabilisticClassifier
-
- setProbabilityCol(String) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- setProbabilityCol(String) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- setProbabilityCol(String) - 类 中的方法org.apache.spark.ml.clustering.GaussianMixture
-
- setProbabilityCol(String) - 类 中的方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- setProductBlocks(int) - 类 中的方法org.apache.spark.mllib.recommendation.ALS
-
Set the number of product blocks to parallelize the computation.
- setPropertiesFile(String) - 类 中的方法org.apache.spark.launcher.AbstractLauncher
-
Set a custom properties file with Spark configuration for the application.
- setPropertiesFile(String) - 类 中的方法org.apache.spark.launcher.SparkLauncher
-
- setQuantileCalculationStrategy(Enumeration.Value) - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- setQuantileProbabilities(double[]) - 类 中的方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- setQuantileProbabilities(double[]) - 类 中的方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- setQuantilesCol(String) - 类 中的方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- setQuantilesCol(String) - 类 中的方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- setRandomCenters(int, double, long) - 类 中的方法org.apache.spark.mllib.clustering.StreamingKMeans
-
Initialize random centers, requiring only the number of dimensions.
- setRank(int) - 类 中的方法org.apache.spark.ml.recommendation.ALS
-
- setRank(int) - 类 中的方法org.apache.spark.mllib.recommendation.ALS
-
Set the rank of the feature matrices computed (number of features).
- setRatingCol(String) - 类 中的方法org.apache.spark.ml.recommendation.ALS
-
- setRawPredictionCol(String) - 类 中的方法org.apache.spark.ml.classification.ClassificationModel
-
- setRawPredictionCol(String) - 类 中的方法org.apache.spark.ml.classification.Classifier
-
- setRawPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- setRawPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- setRawPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- setRawPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- setRawPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- setRawPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- setRawPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- setRawPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- setRawPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- setRawPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- setRawPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- setRawPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- setRawPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- setRawPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- setRawPredictionCol(String) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- setRawPredictionCol(String) - 类 中的方法org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
-
- setRegParam(double) - 类 中的方法org.apache.spark.ml.classification.LinearSVC
-
Set the regularization parameter.
- setRegParam(double) - 类 中的方法org.apache.spark.ml.classification.LogisticRegression
-
Set the regularization parameter.
- setRegParam(double) - 类 中的方法org.apache.spark.ml.recommendation.ALS
-
- setRegParam(double) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
Sets the regularization parameter for L2 regularization.
- setRegParam(double) - 类 中的方法org.apache.spark.ml.regression.LinearRegression
-
Set the regularization parameter.
- setRegParam(double) - 类 中的方法org.apache.spark.mllib.classification.StreamingLogisticRegressionWithSGD
-
Set the regularization parameter.
- setRegParam(double) - 类 中的方法org.apache.spark.mllib.optimization.GradientDescent
-
Set the regularization parameter.
- setRegParam(double) - 类 中的方法org.apache.spark.mllib.optimization.LBFGS
-
Set the regularization parameter.
- setRegParam(double) - 类 中的方法org.apache.spark.mllib.regression.StreamingLinearRegressionWithSGD
-
Set the regularization parameter.
- setRelativeError(double) - 类 中的方法org.apache.spark.ml.feature.QuantileDiscretizer
-
- setRequiredColumns(Configuration, StructType, StructType) - 类 中的静态方法org.apache.spark.sql.hive.orc.OrcFileFormat
-
- setRest(long, int, VD, ED) - 类 中的方法org.apache.spark.graphx.impl.AggregatingEdgeContext
-
- setRuns(int) - 类 中的方法org.apache.spark.mllib.clustering.KMeans
-
- setSample(RDD<Object>) - 类 中的方法org.apache.spark.mllib.stat.KernelDensity
-
Sets the sample to use for density estimation.
- setSample(JavaRDD<Double>) - 类 中的方法org.apache.spark.mllib.stat.KernelDensity
-
Sets the sample to use for density estimation (for Java users).
- setScalingVec(Vector) - 类 中的方法org.apache.spark.ml.feature.ElementwiseProduct
-
- setSeed(long) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- setSeed(long) - 类 中的方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- setSeed(long) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- setSeed(long) - 类 中的方法org.apache.spark.ml.classification.GBTClassifier
-
- setSeed(long) - 类 中的方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
Set the seed for weights initialization if weights are not set
- setSeed(long) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- setSeed(long) - 类 中的方法org.apache.spark.ml.classification.RandomForestClassifier
-
- setSeed(long) - 类 中的方法org.apache.spark.ml.clustering.BisectingKMeans
-
- setSeed(long) - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- setSeed(long) - 类 中的方法org.apache.spark.ml.clustering.GaussianMixture
-
- setSeed(long) - 类 中的方法org.apache.spark.ml.clustering.KMeans
-
- setSeed(long) - 类 中的方法org.apache.spark.ml.clustering.LDA
-
- setSeed(long) - 类 中的方法org.apache.spark.ml.clustering.LDAModel
-
- setSeed(long) - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- setSeed(long) - 类 中的方法org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- setSeed(long) - 类 中的方法org.apache.spark.ml.feature.MinHashLSH
-
- setSeed(long) - 类 中的方法org.apache.spark.ml.feature.Word2Vec
-
- setSeed(long) - 类 中的方法org.apache.spark.ml.recommendation.ALS
-
- setSeed(long) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- setSeed(long) - 类 中的方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- setSeed(long) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- setSeed(long) - 类 中的方法org.apache.spark.ml.regression.GBTRegressor
-
- setSeed(long) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- setSeed(long) - 类 中的方法org.apache.spark.ml.regression.RandomForestRegressor
-
- setSeed(long) - 类 中的方法org.apache.spark.ml.tuning.CrossValidator
-
- setSeed(long) - 类 中的方法org.apache.spark.ml.tuning.TrainValidationSplit
-
- setSeed(long) - 类 中的方法org.apache.spark.mllib.clustering.BisectingKMeans
-
Sets the random seed (default: hash value of the class name).
- setSeed(long) - 类 中的方法org.apache.spark.mllib.clustering.GaussianMixture
-
Set the random seed
- setSeed(long) - 类 中的方法org.apache.spark.mllib.clustering.KMeans
-
Set the random seed for cluster initialization.
- setSeed(long) - 类 中的方法org.apache.spark.mllib.clustering.LDA
-
Set the random seed for cluster initialization.
- setSeed(long) - 类 中的方法org.apache.spark.mllib.feature.Word2Vec
-
Sets random seed (default: a random long integer).
- setSeed(long) - 类 中的方法org.apache.spark.mllib.random.ExponentialGenerator
-
- setSeed(long) - 类 中的方法org.apache.spark.mllib.random.GammaGenerator
-
- setSeed(long) - 类 中的方法org.apache.spark.mllib.random.LogNormalGenerator
-
- setSeed(long) - 类 中的方法org.apache.spark.mllib.random.PoissonGenerator
-
- setSeed(long) - 类 中的方法org.apache.spark.mllib.random.StandardNormalGenerator
-
- setSeed(long) - 类 中的方法org.apache.spark.mllib.random.UniformGenerator
-
- setSeed(long) - 类 中的方法org.apache.spark.mllib.random.WeibullGenerator
-
- setSeed(long) - 类 中的方法org.apache.spark.mllib.recommendation.ALS
-
Sets a random seed to have deterministic results.
- setSeed(long) - 类 中的方法org.apache.spark.util.random.BernoulliCellSampler
-
- setSeed(long) - 类 中的方法org.apache.spark.util.random.BernoulliSampler
-
- setSeed(long) - 类 中的方法org.apache.spark.util.random.PoissonSampler
-
- setSeed(long) - 接口 中的方法org.apache.spark.util.random.Pseudorandom
-
Set random seed.
- setSelectorType(String) - 类 中的方法org.apache.spark.ml.feature.ChiSqSelector
-
- setSelectorType(String) - 类 中的方法org.apache.spark.mllib.feature.ChiSqSelector
-
- setSerializer(Serializer) - 类 中的方法org.apache.spark.rdd.CoGroupedRDD
-
Set a serializer for this RDD's shuffle, or null to use the default (spark.serializer)
- setSerializer(Serializer) - 类 中的方法org.apache.spark.rdd.ShuffledRDD
-
Set a serializer for this RDD's shuffle, or null to use the default (spark.serializer)
- setSize(int) - 类 中的方法org.apache.spark.ml.feature.VectorSizeHint
-
- setSmoothing(double) - 类 中的方法org.apache.spark.ml.classification.NaiveBayes
-
Set the smoothing parameter.
- setSolver(String) - 类 中的方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
Sets the value of param solver.
- setSolver(String) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
Sets the solver algorithm used for optimization.
- setSolver(String) - 类 中的方法org.apache.spark.ml.regression.LinearRegression
-
Set the solver algorithm used for optimization.
- setSparkContextSessionConf(SparkSession, Map<Object, Object>) - 类 中的静态方法org.apache.spark.sql.api.r.SQLUtils
-
- setSparkHome(String) - 类 中的方法org.apache.spark.launcher.SparkLauncher
-
Set a custom Spark installation location for the application.
- setSparkHome(String) - 类 中的方法org.apache.spark.SparkConf
-
Set the location where Spark is installed on worker nodes.
- setSplits(double[]) - 类 中的方法org.apache.spark.ml.feature.Bucketizer
-
- setSplitsArray(double[][]) - 类 中的方法org.apache.spark.ml.feature.Bucketizer
-
- setSQLReadObject(Function2<DataInputStream, Object, Object>) - 类 中的静态方法org.apache.spark.api.r.SerDe
-
- setSQLWriteObject(Function2<DataOutputStream, Object, Object>) - 类 中的静态方法org.apache.spark.api.r.SerDe
-
- setSrcOnly(long, int, VD) - 类 中的方法org.apache.spark.graphx.impl.AggregatingEdgeContext
-
- setStackTrace(StackTraceElement[]) - 异常错误 中的静态方法org.apache.spark.sql.AnalysisException
-
- setStages(PipelineStage[]) - 类 中的方法org.apache.spark.ml.Pipeline
-
- setStandardization(boolean) - 类 中的方法org.apache.spark.ml.classification.LinearSVC
-
Whether to standardize the training features before fitting the model.
- setStandardization(boolean) - 类 中的方法org.apache.spark.ml.classification.LogisticRegression
-
Whether to standardize the training features before fitting the model.
- setStandardization(boolean) - 类 中的方法org.apache.spark.ml.regression.LinearRegression
-
Whether to standardize the training features before fitting the model.
- setStartOffset(Optional<Offset>) - 接口 中的方法org.apache.spark.sql.sources.v2.reader.streaming.ContinuousReader
-
Set the desired start offset for reader factories created from this reader.
- setStatement(String) - 类 中的方法org.apache.spark.ml.feature.SQLTransformer
-
- setStepSize(double) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- setStepSize(double) - 类 中的方法org.apache.spark.ml.classification.GBTClassifier
-
- setStepSize(double) - 类 中的方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
Sets the value of param stepSize (applicable only for solver "gd").
- setStepSize(double) - 类 中的方法org.apache.spark.ml.feature.Word2Vec
-
- setStepSize(double) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- setStepSize(double) - 类 中的方法org.apache.spark.ml.regression.GBTRegressor
-
- setStepSize(double) - 类 中的方法org.apache.spark.mllib.classification.StreamingLogisticRegressionWithSGD
-
Set the step size for gradient descent.
- setStepSize(double) - 类 中的方法org.apache.spark.mllib.optimization.GradientDescent
-
Set the initial step size of SGD for the first step.
- setStepSize(double) - 类 中的方法org.apache.spark.mllib.regression.StreamingLinearRegressionWithSGD
-
Set the step size for gradient descent.
- setStopWords(String[]) - 类 中的方法org.apache.spark.ml.feature.StopWordsRemover
-
- setStorageLevel(String) - 类 中的方法org.apache.spark.status.LiveRDD
-
- setStrategy(String) - 类 中的方法org.apache.spark.ml.feature.Imputer
-
Imputation strategy.
- setStringIndexerOrderType(String) - 类 中的方法org.apache.spark.ml.feature.RFormula
-
- setStringOrderType(String) - 类 中的方法org.apache.spark.ml.feature.StringIndexer
-
- setSubsamplingRate(double) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- setSubsamplingRate(double) - 类 中的方法org.apache.spark.ml.classification.GBTClassifier
-
- setSubsamplingRate(double) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- setSubsamplingRate(double) - 类 中的方法org.apache.spark.ml.classification.RandomForestClassifier
-
- setSubsamplingRate(double) - 类 中的方法org.apache.spark.ml.clustering.LDA
-
- setSubsamplingRate(double) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- setSubsamplingRate(double) - 类 中的方法org.apache.spark.ml.regression.GBTRegressor
-
- setSubsamplingRate(double) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- setSubsamplingRate(double) - 类 中的方法org.apache.spark.ml.regression.RandomForestRegressor
-
- setSubsamplingRate(double) - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- setTau0(double) - 类 中的方法org.apache.spark.mllib.clustering.OnlineLDAOptimizer
-
A (positive) learning parameter that downweights early iterations.
- setTestMethod(String) - 类 中的方法org.apache.spark.mllib.stat.test.StreamingTest
-
Set the statistical method used for significance testing.
- setThreshold(double) - 类 中的方法org.apache.spark.ml.classification.LinearSVC
-
Set threshold in binary classification.
- setThreshold(double) - 类 中的方法org.apache.spark.ml.classification.LinearSVCModel
-
- setThreshold(double) - 类 中的方法org.apache.spark.ml.classification.LogisticRegression
-
- setThreshold(double) - 类 中的方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- setThreshold(double) - 类 中的方法org.apache.spark.ml.feature.Binarizer
-
- setThreshold(double) - 类 中的方法org.apache.spark.mllib.classification.LogisticRegressionModel
-
Sets the threshold that separates positive predictions from negative predictions
in Binary Logistic Regression.
- setThreshold(double) - 类 中的方法org.apache.spark.mllib.classification.SVMModel
-
Sets the threshold that separates positive predictions from negative predictions.
- setThresholds(double[]) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- setThresholds(double[]) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- setThresholds(double[]) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- setThresholds(double[]) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- setThresholds(double[]) - 类 中的方法org.apache.spark.ml.classification.LogisticRegression
-
- setThresholds(double[]) - 类 中的方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- setThresholds(double[]) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- setThresholds(double[]) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- setThresholds(double[]) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- setThresholds(double[]) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- setThresholds(double[]) - 类 中的方法org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- setThresholds(double[]) - 类 中的方法org.apache.spark.ml.classification.ProbabilisticClassifier
-
- setThresholds(double[]) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- setThresholds(double[]) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- setTimeoutDuration(long) - 接口 中的方法org.apache.spark.sql.streaming.GroupState
-
Set the timeout duration in ms for this key.
- setTimeoutDuration(String) - 接口 中的方法org.apache.spark.sql.streaming.GroupState
-
Set the timeout duration for this key as a string.
- setTimeoutTimestamp(long) - 接口 中的方法org.apache.spark.sql.streaming.GroupState
-
Set the timeout timestamp for this key as milliseconds in epoch time.
- setTimeoutTimestamp(long, String) - 接口 中的方法org.apache.spark.sql.streaming.GroupState
-
Set the timeout timestamp for this key as milliseconds in epoch time and an additional
duration as a string (e.g. "1 hour", "2 days", etc.).
- setTimeoutTimestamp(Date) - 接口 中的方法org.apache.spark.sql.streaming.GroupState
-
Set the timeout timestamp for this key as a java.sql.Date.
- setTimeoutTimestamp(Date, String) - 接口 中的方法org.apache.spark.sql.streaming.GroupState
-
Set the timeout timestamp for this key as a java.sql.Date and an additional
duration as a string (e.g. "1 hour", "2 days", etc.).
- setTol(double) - 类 中的方法org.apache.spark.ml.classification.LinearSVC
-
Set the convergence tolerance of iterations.
- setTol(double) - 类 中的方法org.apache.spark.ml.classification.LogisticRegression
-
Set the convergence tolerance of iterations.
- setTol(double) - 类 中的方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
Set the convergence tolerance of iterations.
- setTol(double) - 类 中的方法org.apache.spark.ml.clustering.GaussianMixture
-
- setTol(double) - 类 中的方法org.apache.spark.ml.clustering.KMeans
-
- setTol(double) - 类 中的方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
Set the convergence tolerance of iterations.
- setTol(double) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
Sets the convergence tolerance of iterations.
- setTol(double) - 类 中的方法org.apache.spark.ml.regression.LinearRegression
-
Set the convergence tolerance of iterations.
- setToLowercase(boolean) - 类 中的方法org.apache.spark.ml.feature.RegexTokenizer
-
- setTopicConcentration(double) - 类 中的方法org.apache.spark.ml.clustering.LDA
-
- setTopicConcentration(double) - 类 中的方法org.apache.spark.mllib.clustering.LDA
-
Concentration parameter (commonly named "beta" or "eta") for the prior placed on topics'
distributions over terms.
- setTopicDistributionCol(String) - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- setTopicDistributionCol(String) - 类 中的方法org.apache.spark.ml.clustering.LDA
-
- setTopicDistributionCol(String) - 类 中的方法org.apache.spark.ml.clustering.LDAModel
-
- setTopicDistributionCol(String) - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- setTrainRatio(double) - 类 中的方法org.apache.spark.ml.tuning.TrainValidationSplit
-
- setTreeStrategy(Strategy) - 类 中的方法org.apache.spark.mllib.tree.configuration.BoostingStrategy
-
- setUiRoot(ContextHandler, UIRoot) - 类 中的静态方法org.apache.spark.status.api.v1.UIRootFromServletContext
-
- setupCommitter(TaskAttemptContext) - 类 中的方法org.apache.spark.internal.io.HadoopMapRedCommitProtocol
-
- setUpdater(Updater) - 类 中的方法org.apache.spark.mllib.optimization.GradientDescent
-
Set the updater function to actually perform a gradient step in a given direction.
- setUpdater(Updater) - 类 中的方法org.apache.spark.mllib.optimization.LBFGS
-
Set the updater function to actually perform a gradient step in a given direction.
- SetupDriver(org.apache.spark.rpc.RpcEndpointRef) - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.SetupDriver
-
- SetupDriver$() - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.SetupDriver$
-
- setupGroups(int, DefaultPartitionCoalescer.PartitionLocations) - 类 中的方法org.apache.spark.rdd.DefaultPartitionCoalescer
-
Initializes targetLen partition groups.
- setupJob(JobContext) - 类 中的方法org.apache.spark.internal.io.FileCommitProtocol
-
Setups up a job.
- setupJob(JobContext) - 类 中的方法org.apache.spark.internal.io.HadoopMapReduceCommitProtocol
-
- setUpperBoundsOnCoefficients(Matrix) - 类 中的方法org.apache.spark.ml.classification.LogisticRegression
-
Set the upper bounds on coefficients if fitting under bound constrained optimization.
- setUpperBoundsOnIntercepts(Vector) - 类 中的方法org.apache.spark.ml.classification.LogisticRegression
-
Set the upper bounds on intercepts if fitting under bound constrained optimization.
- setupSecureURLConnection(URLConnection, org.apache.spark.SecurityManager) - 类 中的静态方法org.apache.spark.util.Utils
-
If the given URL connection is HttpsURLConnection, it sets the SSL socket factory and
the host verifier from the given security manager.
- setupTask(TaskAttemptContext) - 类 中的方法org.apache.spark.internal.io.FileCommitProtocol
-
Sets up a task within a job.
- setupTask(TaskAttemptContext) - 类 中的方法org.apache.spark.internal.io.HadoopMapReduceCommitProtocol
-
- setUseNodeIdCache(boolean) - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- setUserBlocks(int) - 类 中的方法org.apache.spark.mllib.recommendation.ALS
-
Set the number of user blocks to parallelize the computation.
- setUserCol(String) - 类 中的方法org.apache.spark.ml.recommendation.ALS
-
- setUserCol(String) - 类 中的方法org.apache.spark.ml.recommendation.ALSModel
-
- setValidateData(boolean) - 类 中的静态方法org.apache.spark.mllib.classification.LogisticRegressionWithSGD
-
- setValidateData(boolean) - 类 中的静态方法org.apache.spark.mllib.classification.SVMWithSGD
-
- setValidateData(boolean) - 类 中的方法org.apache.spark.mllib.regression.GeneralizedLinearAlgorithm
-
Set if the algorithm should validate data before training.
- setValidateData(boolean) - 类 中的静态方法org.apache.spark.mllib.regression.LassoWithSGD
-
- setValidateData(boolean) - 类 中的静态方法org.apache.spark.mllib.regression.LinearRegressionWithSGD
-
- setValidateData(boolean) - 类 中的静态方法org.apache.spark.mllib.regression.RidgeRegressionWithSGD
-
- setValidationTol(double) - 类 中的方法org.apache.spark.mllib.tree.configuration.BoostingStrategy
-
- setValue(R) - 类 中的方法org.apache.spark.Accumulable
-
已过时。
Set the accumulator's value.
- setValue(R) - 类 中的静态方法org.apache.spark.Accumulator
-
已过时。
- setVarianceCol(String) - 类 中的方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- setVarianceCol(String) - 类 中的方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- setVariancePower(double) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
Sets the value of param variancePower.
- setVectorSize(int) - 类 中的方法org.apache.spark.ml.feature.Word2Vec
-
- setVectorSize(int) - 类 中的方法org.apache.spark.mllib.feature.Word2Vec
-
Sets vector size (default: 100).
- setVerbose(boolean) - 类 中的方法org.apache.spark.launcher.AbstractLauncher
-
Enables verbose reporting for SparkSubmit.
- setVerbose(boolean) - 类 中的方法org.apache.spark.launcher.SparkLauncher
-
- setVocabSize(int) - 类 中的方法org.apache.spark.ml.feature.CountVectorizer
-
- setWeightCol(String) - 类 中的方法org.apache.spark.ml.classification.LinearSVC
-
Set the value of param weightCol.
- setWeightCol(double) - 类 中的方法org.apache.spark.ml.classification.LinearSVCModel
-
- setWeightCol(String) - 类 中的方法org.apache.spark.ml.classification.LogisticRegression
-
Sets the value of param weightCol.
- setWeightCol(String) - 类 中的方法org.apache.spark.ml.classification.NaiveBayes
-
Sets the value of param weightCol.
- setWeightCol(String) - 类 中的方法org.apache.spark.ml.classification.OneVsRest
-
Sets the value of param weightCol.
- setWeightCol(String) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
Sets the value of param weightCol.
- setWeightCol(String) - 类 中的方法org.apache.spark.ml.regression.IsotonicRegression
-
- setWeightCol(String) - 类 中的方法org.apache.spark.ml.regression.LinearRegression
-
Whether to over-/under-sample training instances according to the given weights in weightCol.
- setWindowSize(int) - 类 中的方法org.apache.spark.ml.feature.Word2Vec
-
- setWindowSize(int) - 类 中的方法org.apache.spark.mllib.feature.Word2Vec
-
Sets the window of words (default: 5)
- setWindowSize(int) - 类 中的方法org.apache.spark.mllib.stat.test.StreamingTest
-
Set the number of batches to compute significance tests over.
- setWithMean(boolean) - 类 中的方法org.apache.spark.ml.feature.StandardScaler
-
- setWithMean(boolean) - 类 中的方法org.apache.spark.mllib.feature.StandardScalerModel
-
:: DeveloperApi ::
- setWithStd(boolean) - 类 中的方法org.apache.spark.ml.feature.StandardScaler
-
- setWithStd(boolean) - 类 中的方法org.apache.spark.mllib.feature.StandardScalerModel
-
:: DeveloperApi ::
- sha1(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Calculates the SHA-1 digest of a binary column and returns the value
as a 40 character hex string.
- sha2(Column, int) - 类 中的静态方法org.apache.spark.sql.functions
-
Calculates the SHA-2 family of hash functions of a binary column and
returns the value as a hex string.
- shape() - 类 中的方法org.apache.spark.mllib.random.GammaGenerator
-
- SharedParamsCodeGen - org.apache.spark.ml.param.shared中的类
-
Code generator for shared params (sharedParams.scala).
- SharedParamsCodeGen() - 类 的构造器org.apache.spark.ml.param.shared.SharedParamsCodeGen
-
- SharedReadWrite$() - 类 的构造器org.apache.spark.ml.Pipeline.SharedReadWrite$
-
- sharedState() - 类 中的方法org.apache.spark.sql.SparkSession
-
State shared across sessions, including the SparkContext, cached data, listener,
and a catalog that interacts with external systems.
- shiftLeft(Column, int) - 类 中的静态方法org.apache.spark.sql.functions
-
Shift the given value numBits left.
- shiftRight(Column, int) - 类 中的静态方法org.apache.spark.sql.functions
-
(Signed) shift the given value numBits right.
- shiftRightUnsigned(Column, int) - 类 中的静态方法org.apache.spark.sql.functions
-
Unsigned shift the given value numBits right.
- SHORT() - 类 中的静态方法org.apache.spark.sql.Encoders
-
An encoder for nullable short type.
- ShortestPaths - org.apache.spark.graphx.lib中的类
-
Computes shortest paths to the given set of landmark vertices, returning a graph where each
vertex attribute is a map containing the shortest-path distance to each reachable landmark.
- ShortestPaths() - 类 的构造器org.apache.spark.graphx.lib.ShortestPaths
-
- shortName() - 类 中的方法org.apache.spark.sql.hive.execution.HiveFileFormat
-
- shortName() - 类 中的方法org.apache.spark.sql.hive.orc.OrcFileFormat
-
- shortName() - 接口 中的方法org.apache.spark.sql.sources.DataSourceRegister
-
The string that represents the format that this data source provider uses.
- shortTimeUnitString(TimeUnit) - 类 中的静态方法org.apache.spark.streaming.ui.UIUtils
-
Return the short string for a TimeUnit.
- ShortType - 类 中的静态变量org.apache.spark.sql.types.DataTypes
-
Gets the ShortType object.
- ShortType - org.apache.spark.sql.types中的类
-
The data type representing Short values.
- shouldCloseFileAfterWrite(SparkConf, boolean) - 类 中的静态方法org.apache.spark.streaming.util.WriteAheadLogUtils
-
- shouldDistributeGaussians(int, int) - 类 中的静态方法org.apache.spark.mllib.clustering.GaussianMixture
-
Heuristic to distribute the computation of the MultivariateGaussians, approximately when
d is greater than 25 except for when k is very small.
- shouldGoLeft(Vector) - 接口 中的方法org.apache.spark.ml.tree.Split
-
Return true (split to left) or false (split to right).
- shouldGoLeft(int, Split[]) - 接口 中的方法org.apache.spark.ml.tree.Split
-
Return true (split to left) or false (split to right).
- shouldOwn(Param<?>) - 接口 中的方法org.apache.spark.ml.param.Params
-
Validates that the input param belongs to this instance.
- show(int) - 类 中的方法org.apache.spark.sql.Dataset
-
Displays the Dataset in a tabular form.
- show() - 类 中的方法org.apache.spark.sql.Dataset
-
Displays the top 20 rows of Dataset in a tabular form.
- show(boolean) - 类 中的方法org.apache.spark.sql.Dataset
-
Displays the top 20 rows of Dataset in a tabular form.
- show(int, boolean) - 类 中的方法org.apache.spark.sql.Dataset
-
Displays the Dataset in a tabular form.
- show(int, int) - 类 中的方法org.apache.spark.sql.Dataset
-
Displays the Dataset in a tabular form.
- show(int, int, boolean) - 类 中的方法org.apache.spark.sql.Dataset
-
Displays the Dataset in a tabular form.
- showBytesDistribution(String, Function2<TaskInfo, TaskMetrics, Object>, Seq<Tuple2<TaskInfo, TaskMetrics>>) - 类 中的静态方法org.apache.spark.scheduler.StatsReportListener
-
- showBytesDistribution(String, Option<org.apache.spark.util.Distribution>) - 类 中的静态方法org.apache.spark.scheduler.StatsReportListener
-
- showBytesDistribution(String, org.apache.spark.util.Distribution) - 类 中的静态方法org.apache.spark.scheduler.StatsReportListener
-
- showDagVizForJob(int, Seq<org.apache.spark.ui.scope.RDDOperationGraph>) - 类 中的静态方法org.apache.spark.ui.UIUtils
-
Return a "DAG visualization" DOM element that expands into a visualization for a job.
- showDagVizForStage(int, Option<org.apache.spark.ui.scope.RDDOperationGraph>) - 类 中的静态方法org.apache.spark.ui.UIUtils
-
Return a "DAG visualization" DOM element that expands into a visualization for a stage.
- showDistribution(String, org.apache.spark.util.Distribution, Function1<Object, String>) - 类 中的静态方法org.apache.spark.scheduler.StatsReportListener
-
- showDistribution(String, Option<org.apache.spark.util.Distribution>, Function1<Object, String>) - 类 中的静态方法org.apache.spark.scheduler.StatsReportListener
-
- showDistribution(String, Option<org.apache.spark.util.Distribution>, String) - 类 中的静态方法org.apache.spark.scheduler.StatsReportListener
-
- showDistribution(String, String, Function2<TaskInfo, TaskMetrics, Object>, Seq<Tuple2<TaskInfo, TaskMetrics>>) - 类 中的静态方法org.apache.spark.scheduler.StatsReportListener
-
- showMillisDistribution(String, Option<org.apache.spark.util.Distribution>) - 类 中的静态方法org.apache.spark.scheduler.StatsReportListener
-
- showMillisDistribution(String, Function2<TaskInfo, TaskMetrics, Object>, Seq<Tuple2<TaskInfo, TaskMetrics>>) - 类 中的静态方法org.apache.spark.scheduler.StatsReportListener
-
- showMillisDistribution(String, Function1<BatchInfo, Option<Object>>) - 类 中的方法org.apache.spark.streaming.scheduler.StatsReportListener
-
- SHUFFLE() - 类 中的静态方法org.apache.spark.storage.BlockId
-
- SHUFFLE_DATA() - 类 中的静态方法org.apache.spark.storage.BlockId
-
- SHUFFLE_INDEX() - 类 中的静态方法org.apache.spark.storage.BlockId
-
- SHUFFLE_LOCAL_BLOCKS() - 类 中的静态方法org.apache.spark.status.TaskIndexNames
-
- SHUFFLE_READ() - 类 中的静态方法org.apache.spark.ui.ToolTips
-
- SHUFFLE_READ_BLOCKED_TIME() - 类 中的静态方法org.apache.spark.ui.jobs.TaskDetailsClassNames
-
- SHUFFLE_READ_BLOCKED_TIME() - 类 中的静态方法org.apache.spark.ui.ToolTips
-
- SHUFFLE_READ_METRICS_PREFIX() - 类 中的静态方法org.apache.spark.InternalAccumulator
-
- SHUFFLE_READ_RECORDS() - 类 中的静态方法org.apache.spark.status.TaskIndexNames
-
- SHUFFLE_READ_REMOTE_SIZE() - 类 中的静态方法org.apache.spark.ui.jobs.TaskDetailsClassNames
-
- SHUFFLE_READ_REMOTE_SIZE() - 类 中的静态方法org.apache.spark.ui.ToolTips
-
- SHUFFLE_READ_TIME() - 类 中的静态方法org.apache.spark.status.TaskIndexNames
-
- SHUFFLE_REMOTE_BLOCKS() - 类 中的静态方法org.apache.spark.status.TaskIndexNames
-
- SHUFFLE_REMOTE_READS() - 类 中的静态方法org.apache.spark.status.TaskIndexNames
-
- SHUFFLE_REMOTE_READS_TO_DISK() - 类 中的静态方法org.apache.spark.status.TaskIndexNames
-
- SHUFFLE_TOTAL_BLOCKS() - 类 中的静态方法org.apache.spark.status.TaskIndexNames
-
- SHUFFLE_TOTAL_READS() - 类 中的静态方法org.apache.spark.status.TaskIndexNames
-
- SHUFFLE_WRITE() - 类 中的静态方法org.apache.spark.ui.ToolTips
-
- SHUFFLE_WRITE_METRICS_PREFIX() - 类 中的静态方法org.apache.spark.InternalAccumulator
-
- SHUFFLE_WRITE_RECORDS() - 类 中的静态方法org.apache.spark.status.TaskIndexNames
-
- SHUFFLE_WRITE_SIZE() - 类 中的静态方法org.apache.spark.status.TaskIndexNames
-
- SHUFFLE_WRITE_TIME() - 类 中的静态方法org.apache.spark.status.TaskIndexNames
-
- ShuffleBlockId - org.apache.spark.storage中的类
-
- ShuffleBlockId(int, int, int) - 类 的构造器org.apache.spark.storage.ShuffleBlockId
-
- ShuffleDataBlockId - org.apache.spark.storage中的类
-
- ShuffleDataBlockId(int, int, int) - 类 的构造器org.apache.spark.storage.ShuffleDataBlockId
-
- ShuffleDependency<K,V,C> - org.apache.spark中的类
-
:: DeveloperApi ::
Represents a dependency on the output of a shuffle stage.
- ShuffleDependency(RDD<? extends Product2<K, V>>, Partitioner, Serializer, Option<Ordering<K>>, Option<Aggregator<K, V, C>>, boolean, ClassTag<K>, ClassTag<V>, ClassTag<C>) - 类 的构造器org.apache.spark.ShuffleDependency
-
- ShuffledRDD<K,V,C> - org.apache.spark.rdd中的类
-
:: DeveloperApi ::
The resulting RDD from a shuffle (e.g. repartitioning of data).
- ShuffledRDD(RDD<? extends Product2<K, V>>, Partitioner, ClassTag<K>, ClassTag<V>, ClassTag<C>) - 类 的构造器org.apache.spark.rdd.ShuffledRDD
-
- shuffleHandle() - 类 中的方法org.apache.spark.ShuffleDependency
-
- shuffleId() - 类 中的方法org.apache.spark.CleanShuffle
-
- shuffleId() - 类 中的方法org.apache.spark.FetchFailed
-
- shuffleId() - 类 中的方法org.apache.spark.ShuffleDependency
-
- shuffleId() - 类 中的方法org.apache.spark.storage.BlockManagerMessages.RemoveShuffle
-
- shuffleId() - 类 中的方法org.apache.spark.storage.ShuffleBlockId
-
- shuffleId() - 类 中的方法org.apache.spark.storage.ShuffleDataBlockId
-
- shuffleId() - 类 中的方法org.apache.spark.storage.ShuffleIndexBlockId
-
- ShuffleIndexBlockId - org.apache.spark.storage中的类
-
- ShuffleIndexBlockId(int, int, int) - 类 的构造器org.apache.spark.storage.ShuffleIndexBlockId
-
- shuffleManager() - 类 中的方法org.apache.spark.SparkEnv
-
- shuffleRead() - 类 中的方法org.apache.spark.status.api.v1.ExecutorStageSummary
-
- shuffleRead$() - 类 的构造器org.apache.spark.InternalAccumulator.shuffleRead$
-
- shuffleReadBytes() - 类 中的方法org.apache.spark.status.api.v1.StageData
-
- ShuffleReadMetricDistributions - org.apache.spark.status.api.v1中的类
-
- ShuffleReadMetrics - org.apache.spark.status.api.v1中的类
-
- shuffleReadMetrics() - 类 中的方法org.apache.spark.status.api.v1.TaskMetricDistributions
-
- shuffleReadMetrics() - 类 中的方法org.apache.spark.status.api.v1.TaskMetrics
-
- shuffleReadRecords() - 类 中的方法org.apache.spark.status.api.v1.ExecutorStageSummary
-
- shuffleReadRecords() - 类 中的方法org.apache.spark.status.api.v1.StageData
-
- ShuffleStatus - org.apache.spark中的类
-
Helper class used by the MapOutputTrackerMaster to perform bookkeeping for a single
ShuffleMapStage.
- ShuffleStatus(int) - 类 的构造器org.apache.spark.ShuffleStatus
-
- shuffleWrite() - 类 中的方法org.apache.spark.status.api.v1.ExecutorStageSummary
-
- shuffleWrite$() - 类 的构造器org.apache.spark.InternalAccumulator.shuffleWrite$
-
- shuffleWriteBytes() - 类 中的方法org.apache.spark.status.api.v1.StageData
-
- ShuffleWriteMetricDistributions - org.apache.spark.status.api.v1中的类
-
- ShuffleWriteMetrics - org.apache.spark.status.api.v1中的类
-
- shuffleWriteMetrics() - 类 中的方法org.apache.spark.status.api.v1.TaskMetricDistributions
-
- shuffleWriteMetrics() - 类 中的方法org.apache.spark.status.api.v1.TaskMetrics
-
- shuffleWriteRecords() - 类 中的方法org.apache.spark.status.api.v1.ExecutorStageSummary
-
- shuffleWriteRecords() - 类 中的方法org.apache.spark.status.api.v1.StageData
-
- Shutdown$() - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.Shutdown$
-
- ShutdownHookManager - org.apache.spark.util中的类
-
Various utility methods used by Spark.
- ShutdownHookManager() - 类 的构造器org.apache.spark.util.ShutdownHookManager
-
- sigma() - 类 中的方法org.apache.spark.mllib.stat.distribution.MultivariateGaussian
-
- sigmas() - 类 中的方法org.apache.spark.mllib.clustering.ExpectationSum
-
- SignalUtils - org.apache.spark.util中的类
-
Contains utilities for working with posix signals.
- SignalUtils() - 类 的构造器org.apache.spark.util.SignalUtils
-
- signum(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Computes the signum of the given value.
- signum(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Computes the signum of the given column.
- SimpleFutureAction<T> - org.apache.spark中的类
-
A
FutureAction holding the result of an action that triggers a single job.
- simpleString() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- simpleString() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- simpleString() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- simpleString() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- simpleString() - 类 中的方法org.apache.spark.sql.types.ArrayType
-
- simpleString() - 类 中的静态方法org.apache.spark.sql.types.BinaryType
-
- simpleString() - 类 中的静态方法org.apache.spark.sql.types.BooleanType
-
- simpleString() - 类 中的方法org.apache.spark.sql.types.ByteType
-
- simpleString() - 类 中的静态方法org.apache.spark.sql.types.CalendarIntervalType
-
- simpleString() - 类 中的方法org.apache.spark.sql.types.CharType
-
- simpleString() - 类 中的方法org.apache.spark.sql.types.DataType
-
Readable string representation for the type.
- simpleString() - 类 中的静态方法org.apache.spark.sql.types.DateType
-
- simpleString() - 类 中的方法org.apache.spark.sql.types.DecimalType
-
- simpleString() - 类 中的静态方法org.apache.spark.sql.types.DoubleType
-
- simpleString() - 类 中的静态方法org.apache.spark.sql.types.FloatType
-
- simpleString() - 类 中的静态方法org.apache.spark.sql.types.HiveStringType
-
- simpleString() - 类 中的方法org.apache.spark.sql.types.IntegerType
-
- simpleString() - 类 中的方法org.apache.spark.sql.types.LongType
-
- simpleString() - 类 中的方法org.apache.spark.sql.types.MapType
-
- simpleString() - 类 中的静态方法org.apache.spark.sql.types.NullType
-
- simpleString() - 类 中的方法org.apache.spark.sql.types.ObjectType
-
- simpleString() - 类 中的方法org.apache.spark.sql.types.ShortType
-
- simpleString() - 类 中的静态方法org.apache.spark.sql.types.StringType
-
- simpleString() - 类 中的方法org.apache.spark.sql.types.StructType
-
- simpleString() - 类 中的静态方法org.apache.spark.sql.types.TimestampType
-
- simpleString() - 类 中的方法org.apache.spark.sql.types.VarcharType
-
- SimpleUpdater - org.apache.spark.mllib.optimization中的类
-
:: DeveloperApi ::
A simple updater for gradient descent *without* any regularization.
- SimpleUpdater() - 类 的构造器org.apache.spark.mllib.optimization.SimpleUpdater
-
- sin(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
- sin(String) - 类 中的静态方法org.apache.spark.sql.functions
-
- SingularValueDecomposition<UType,VType> - org.apache.spark.mllib.linalg中的类
-
Represents singular value decomposition (SVD) factors.
- SingularValueDecomposition(UType, Vector, VType) - 类 的构造器org.apache.spark.mllib.linalg.SingularValueDecomposition
-
- sinh(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
- sinh(String) - 类 中的静态方法org.apache.spark.sql.functions
-
- sink() - 类 中的方法org.apache.spark.sql.streaming.StreamingQueryProgress
-
- SinkProgress - org.apache.spark.sql.streaming中的类
-
Information about progress made for a sink in the execution of a
StreamingQuery
during a trigger.
- size() - 类 中的方法org.apache.spark.api.java.JavaUtils.SerializableMapWrapper
-
- size() - 类 中的方法org.apache.spark.ml.attribute.AttributeGroup
-
Size of the attribute group.
- size() - 类 中的方法org.apache.spark.ml.feature.VectorSizeHint
-
The size of Vectors in inputCol.
- size() - 类 中的方法org.apache.spark.ml.linalg.DenseVector
-
- size() - 类 中的方法org.apache.spark.ml.linalg.SparseVector
-
- size() - 接口 中的方法org.apache.spark.ml.linalg.Vector
-
Size of the vector.
- size() - 类 中的方法org.apache.spark.ml.param.ParamMap
-
Number of param pairs in this map.
- size() - 类 中的方法org.apache.spark.mllib.linalg.DenseVector
-
- size() - 类 中的方法org.apache.spark.mllib.linalg.SparseVector
-
- size() - 接口 中的方法org.apache.spark.mllib.linalg.Vector
-
Size of the vector.
- size(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns length of array or map.
- size() - 接口 中的方法org.apache.spark.sql.Row
-
Number of elements in the Row.
- size() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- size() - 类 中的方法org.apache.spark.storage.DiskBlockData
-
- size() - 类 中的方法org.apache.spark.storage.EncryptedBlockData
-
- size() - 类 中的方法org.apache.spark.storage.memory.DeserializedMemoryEntry
-
- size() - 接口 中的方法org.apache.spark.storage.memory.MemoryEntry
-
- size() - 类 中的方法org.apache.spark.storage.memory.SerializedMemoryEntry
-
- SizeEstimator - org.apache.spark.util中的类
-
:: DeveloperApi ::
Estimates the sizes of Java objects (number of bytes of memory they occupy), for use in
memory-aware caches.
- SizeEstimator() - 类 的构造器org.apache.spark.util.SizeEstimator
-
- sizeInBytes() - 类 中的方法org.apache.spark.sql.sources.BaseRelation
-
Returns an estimated size of this relation in bytes.
- sizeInBytes() - 接口 中的方法org.apache.spark.sql.sources.v2.reader.Statistics
-
- sketch(RDD<K>, int, ClassTag<K>) - 类 中的静态方法org.apache.spark.RangePartitioner
-
Sketches the input RDD via reservoir sampling on each partition.
- skewness(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the skewness of the values in a group.
- skewness(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the skewness of the values in a group.
- skip(long) - 类 中的方法org.apache.spark.io.NioBufferedFileInputStream
-
- skip(long) - 类 中的方法org.apache.spark.io.ReadAheadInputStream
-
- skip(long) - 类 中的方法org.apache.spark.storage.BufferReleasingInputStream
-
- skip(long) - 接口 中的方法org.apache.spark.util.kvstore.KVStoreIterator
-
Skip in the iterator.
- skip(long) - 类 中的方法org.apache.spark.util.kvstore.KVStoreView
-
Skips a number of elements at the start of iteration.
- skippedStages() - 类 中的方法org.apache.spark.status.LiveJob
-
- skippedTasks() - 类 中的方法org.apache.spark.status.LiveJob
-
- skipWhitespace() - 类 中的静态方法org.apache.spark.ml.feature.RFormulaParser
-
- slice(int, int) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- slice(Time, Time) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- slice(Time, Time) - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return all the RDDs between 'fromDuration' to 'toDuration' (both included)
- slice(Time, Time) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- slice(Time, Time) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- slice(Time, Time) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- slice(Time, Time) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- slice(Time, Time) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- slice(org.apache.spark.streaming.Interval) - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Return all the RDDs defined by the Interval object (both end times included)
- slice(Time, Time) - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Return all the RDDs between 'fromTime' to 'toTime' (both included)
- slideDuration() - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Time interval after which the DStream generates an RDD
- slideDuration() - 类 中的方法org.apache.spark.streaming.dstream.InputDStream
-
- sliding(int, int) - 类 中的方法org.apache.spark.mllib.rdd.RDDFunctions
-
Returns an RDD from grouping items of its parent RDD in fixed size blocks by passing a sliding
window over them.
- sliding(int) - 类 中的方法org.apache.spark.mllib.rdd.RDDFunctions
-
sliding(Int, Int)* with step = 1.
- sliding(int) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- sliding(int, int) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- smoothing() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- smoothing() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- SnappyCompressionCodec - org.apache.spark.io中的类
-
- SnappyCompressionCodec(SparkConf) - 类 的构造器org.apache.spark.io.SnappyCompressionCodec
-
- SnappyOutputStreamWrapper - org.apache.spark.io中的类
-
Wrapper over SnappyOutputStream which guards against write-after-close and double-close
issues.
- SnappyOutputStreamWrapper(SnappyOutputStream) - 类 的构造器org.apache.spark.io.SnappyOutputStreamWrapper
-
- socketStream(String, int, Function<InputStream, Iterable<T>>, StorageLevel) - 类 中的方法org.apache.spark.streaming.api.java.JavaStreamingContext
-
Create an input stream from network source hostname:port.
- socketStream(String, int, Function1<InputStream, Iterator<T>>, StorageLevel, ClassTag<T>) - 类 中的方法org.apache.spark.streaming.StreamingContext
-
Creates an input stream from TCP source hostname:port.
- socketTextStream(String, int, StorageLevel) - 类 中的方法org.apache.spark.streaming.api.java.JavaStreamingContext
-
Create an input stream from network source hostname:port.
- socketTextStream(String, int) - 类 中的方法org.apache.spark.streaming.api.java.JavaStreamingContext
-
Create an input stream from network source hostname:port.
- socketTextStream(String, int, StorageLevel) - 类 中的方法org.apache.spark.streaming.StreamingContext
-
Creates an input stream from TCP source hostname:port.
- solve(double[], double[]) - 类 中的静态方法org.apache.spark.mllib.linalg.CholeskyDecomposition
-
Solves a symmetric positive definite linear system via Cholesky factorization.
- solve(double[], double[], NNLS.Workspace) - 类 中的静态方法org.apache.spark.mllib.optimization.NNLS
-
Solve a least squares problem, possibly with nonnegativity constraints, by a modified
projected gradient method.
- solver() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- solver() - 接口 中的方法org.apache.spark.ml.param.shared.HasSolver
-
Param for the solver algorithm for optimization.
- solver() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- solver() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- solver() - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegressionTrainingSummary
-
- solver() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- solver() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- Sort() - 类 中的静态方法org.apache.spark.mllib.tree.configuration.QuantileStrategy
-
- sort(String, String...) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new Dataset sorted by the specified column, all in ascending order.
- sort(Column...) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new Dataset sorted by the given expressions.
- sort(String, Seq<String>) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new Dataset sorted by the specified column, all in ascending order.
- sort(Seq<Column>) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new Dataset sorted by the given expressions.
- sort_array(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Sorts the input array for the given column in ascending order,
according to the natural ordering of the array elements.
- sort_array(Column, boolean) - 类 中的静态方法org.apache.spark.sql.functions
-
Sorts the input array for the given column in ascending or descending order,
according to the natural ordering of the array elements.
- sortBy(Function<T, S>, boolean, int) - 类 中的方法org.apache.spark.api.java.JavaRDD
-
Return this RDD sorted by the given key function.
- sortBy(Function1<T, K>, boolean, int, Ordering<K>, ClassTag<K>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- sortBy(Function1<T, K>, boolean, int, Ordering<K>, ClassTag<K>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- sortBy(Function1<T, K>, boolean, int, Ordering<K>, ClassTag<K>) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- sortBy(Function1<T, K>, boolean, int, Ordering<K>, ClassTag<K>) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- sortBy(Function1<T, K>, boolean, int, Ordering<K>, ClassTag<K>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- sortBy(Function1<T, K>, boolean, int, Ordering<K>, ClassTag<K>) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- sortBy(Function1<T, K>, boolean, int, Ordering<K>, ClassTag<K>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- sortBy(Function1<T, K>, boolean, int, Ordering<K>, ClassTag<K>) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- sortBy(Function1<T, K>, boolean, int, Ordering<K>, ClassTag<K>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- sortBy(Function1<T, K>, boolean, int, Ordering<K>, ClassTag<K>) - 类 中的方法org.apache.spark.rdd.RDD
-
Return this RDD sorted by the given key function.
- sortBy(Function1<T, K>, boolean, int, Ordering<K>, ClassTag<K>) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- sortBy(String, String...) - 类 中的方法org.apache.spark.sql.DataFrameWriter
-
Sorts the output in each bucket by the given columns.
- sortBy(String, Seq<String>) - 类 中的方法org.apache.spark.sql.DataFrameWriter
-
Sorts the output in each bucket by the given columns.
- sortBy(Function1<A, B>, Ordering<B>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- sortBy$default$2() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- sortBy$default$2() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- sortBy$default$2() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- sortBy$default$2() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- sortBy$default$2() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- sortBy$default$2() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- sortBy$default$2() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- sortBy$default$2() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- sortBy$default$2() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- sortBy$default$2() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- sortBy$default$3() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- sortBy$default$3() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- sortBy$default$3() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- sortBy$default$3() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- sortBy$default$3() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- sortBy$default$3() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- sortBy$default$3() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- sortBy$default$3() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- sortBy$default$3() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- sortBy$default$3() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- sortByKey() - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Sort the RDD by key, so that each partition contains a sorted range of the elements in
ascending order.
- sortByKey(boolean) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Sort the RDD by key, so that each partition contains a sorted range of the elements.
- sortByKey(boolean, int) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Sort the RDD by key, so that each partition contains a sorted range of the elements.
- sortByKey(Comparator<K>) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Sort the RDD by key, so that each partition contains a sorted range of the elements.
- sortByKey(Comparator<K>, boolean) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Sort the RDD by key, so that each partition contains a sorted range of the elements.
- sortByKey(Comparator<K>, boolean, int) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Sort the RDD by key, so that each partition contains a sorted range of the elements.
- sortByKey(boolean, int) - 类 中的方法org.apache.spark.rdd.OrderedRDDFunctions
-
Sort the RDD by key, so that each partition contains a sorted range of the elements.
- sorted(Ordering<B>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- sortWith(Function2<A, A, Object>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- sortWithinPartitions(String, String...) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new Dataset with each partition sorted by the given expressions.
- sortWithinPartitions(Column...) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new Dataset with each partition sorted by the given expressions.
- sortWithinPartitions(String, Seq<String>) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new Dataset with each partition sorted by the given expressions.
- sortWithinPartitions(Seq<Column>) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns a new Dataset with each partition sorted by the given expressions.
- soundex(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the soundex code for the specified expression.
- sourceName() - 类 中的静态方法org.apache.spark.metrics.source.CodegenMetrics
-
- sourceName() - 类 中的静态方法org.apache.spark.metrics.source.HiveCatalogMetrics
-
- SourceProgress - org.apache.spark.sql.streaming中的类
-
Information about progress made for a source in the execution of a
StreamingQuery
during a trigger.
- sources() - 类 中的方法org.apache.spark.sql.streaming.StreamingQueryProgress
-
- sourceSchema(SQLContext, Option<StructType>, String, Map<String, String>) - 接口 中的方法org.apache.spark.sql.sources.StreamSourceProvider
-
Returns the name and schema of the source that can be used to continually read data.
- span(Function1<A, Object>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- spark() - 类 中的方法org.apache.spark.status.api.v1.VersionInfo
-
- SPARK_CONNECTOR_NAME() - 类 中的静态方法org.apache.spark.ui.JettyUtils
-
- SPARK_CONTEXT_SHUTDOWN_PRIORITY() - 类 中的静态方法org.apache.spark.util.ShutdownHookManager
-
The shutdown priority of the SparkContext instance.
- SPARK_IO_ENCRYPTION_COMMONS_CONFIG_PREFIX() - 类 中的静态方法org.apache.spark.security.CryptoStreamUtils
-
- SPARK_MASTER - 类 中的静态变量org.apache.spark.launcher.SparkLauncher
-
The Spark master.
- spark_partition_id() - 类 中的静态方法org.apache.spark.sql.functions
-
Partition ID.
- SPARK_REGEX() - 类 中的静态方法org.apache.spark.SparkMasterRegex
-
- SparkAppConfig(Seq<Tuple2<String, String>>, Option<byte[]>, Option<byte[]>) - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.SparkAppConfig
-
- SparkAppConfig$() - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.SparkAppConfig$
-
- SparkAppHandle - org.apache.spark.launcher中的接口
-
A handle to a running Spark application.
- SparkAppHandle.Listener - org.apache.spark.launcher中的接口
-
Listener for updates to a handle's state.
- SparkAppHandle.State - org.apache.spark.launcher中的枚举
-
Represents the application's state.
- SparkConf - org.apache.spark中的类
-
Configuration for a Spark application.
- SparkConf(boolean) - 类 的构造器org.apache.spark.SparkConf
-
- SparkConf() - 类 的构造器org.apache.spark.SparkConf
-
Create a SparkConf that loads defaults from system properties and the classpath
- sparkContext() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- sparkContext() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- sparkContext() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- sparkContext() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- sparkContext() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- sparkContext() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- sparkContext() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- sparkContext() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- sparkContext() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- sparkContext() - 类 中的方法org.apache.spark.rdd.RDD
-
The SparkContext that created this RDD.
- sparkContext() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- SparkContext - org.apache.spark中的类
-
Main entry point for Spark functionality.
- SparkContext(SparkConf) - 类 的构造器org.apache.spark.SparkContext
-
- SparkContext() - 类 的构造器org.apache.spark.SparkContext
-
Create a SparkContext that loads settings from system properties (for instance, when
launching with .
- SparkContext(String, String, SparkConf) - 类 的构造器org.apache.spark.SparkContext
-
Alternative constructor that allows setting common Spark properties directly
- SparkContext(String, String, String, Seq<String>, Map<String, String>) - 类 的构造器org.apache.spark.SparkContext
-
Alternative constructor that allows setting common Spark properties directly
- sparkContext() - 类 中的方法org.apache.spark.sql.SparkSession
-
- sparkContext() - 类 中的方法org.apache.spark.sql.SQLContext
-
- sparkContext() - 类 中的方法org.apache.spark.streaming.api.java.JavaStreamingContext
-
The underlying SparkContext
- sparkContext() - 类 中的方法org.apache.spark.streaming.StreamingContext
-
Return the associated Spark context
- SparkEnv - org.apache.spark中的类
-
:: DeveloperApi ::
Holds all the runtime environment objects for a running Spark instance (either master or worker),
including the serializer, RpcEnv, block manager, map output tracker, etc.
- SparkEnv(String, org.apache.spark.rpc.RpcEnv, Serializer, Serializer, SerializerManager, MapOutputTracker, ShuffleManager, org.apache.spark.broadcast.BroadcastManager, BlockManager, SecurityManager, org.apache.spark.metrics.MetricsSystem, MemoryManager, OutputCommitCoordinator, SparkConf) - 类 的构造器org.apache.spark.SparkEnv
-
- sparkEventFromJson(JsonAST.JValue) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- sparkEventToJson(SparkListenerEvent) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
------------------------------------------------- *
JSON serialization methods for SparkListenerEvents |
- SparkException - org.apache.spark中的异常错误
-
- SparkException(String, Throwable) - 异常错误 的构造器org.apache.spark.SparkException
-
- SparkException(String) - 异常错误 的构造器org.apache.spark.SparkException
-
- SparkExecutorInfo - org.apache.spark中的接口
-
Exposes information about Spark Executors.
- SparkExecutorInfoImpl - org.apache.spark中的类
-
- SparkExecutorInfoImpl(String, int, long, int) - 类 的构造器org.apache.spark.SparkExecutorInfoImpl
-
- SparkExitCode - org.apache.spark.util中的类
-
- SparkExitCode() - 类 的构造器org.apache.spark.util.SparkExitCode
-
- SparkFiles - org.apache.spark中的类
-
Resolves paths to files added through SparkContext.addFile().
- SparkFiles() - 类 的构造器org.apache.spark.SparkFiles
-
- SparkFirehoseListener - org.apache.spark中的类
-
Class that allows users to receive all SparkListener events.
- SparkFirehoseListener() - 类 的构造器org.apache.spark.SparkFirehoseListener
-
- SparkHadoopMapRedUtil - org.apache.spark.mapred中的类
-
- SparkHadoopMapRedUtil() - 类 的构造器org.apache.spark.mapred.SparkHadoopMapRedUtil
-
- SparkHadoopWriter - org.apache.spark.internal.io中的类
-
A helper object that saves an RDD using a Hadoop OutputFormat.
- SparkHadoopWriter() - 类 的构造器org.apache.spark.internal.io.SparkHadoopWriter
-
- SparkHadoopWriterUtils - org.apache.spark.internal.io中的类
-
A helper object that provide common utils used during saving an RDD using a Hadoop OutputFormat
(both from the old mapred API and the new mapreduce API)
- SparkHadoopWriterUtils() - 类 的构造器org.apache.spark.internal.io.SparkHadoopWriterUtils
-
- sparkJavaOpts(SparkConf, Function1<String, Object>) - 类 中的静态方法org.apache.spark.util.Utils
-
Convert all spark properties set in the given SparkConf to a sequence of java options.
- SparkJobInfo - org.apache.spark中的接口
-
Exposes information about Spark Jobs.
- SparkJobInfoImpl - org.apache.spark中的类
-
- SparkJobInfoImpl(int, int[], JobExecutionStatus) - 类 的构造器org.apache.spark.SparkJobInfoImpl
-
- SparkLauncher - org.apache.spark.launcher中的类
-
Launcher for Spark applications.
- SparkLauncher() - 类 的构造器org.apache.spark.launcher.SparkLauncher
-
- SparkLauncher(Map<String, String>) - 类 的构造器org.apache.spark.launcher.SparkLauncher
-
Creates a launcher that will set the given environment variables in the child.
- SparkListener - org.apache.spark.scheduler中的类
-
:: DeveloperApi ::
A default implementation for SparkListenerInterface that has no-op implementations for
all callbacks.
- SparkListener() - 类 的构造器org.apache.spark.scheduler.SparkListener
-
- SparkListenerApplicationEnd - org.apache.spark.scheduler中的类
-
- SparkListenerApplicationEnd(long) - 类 的构造器org.apache.spark.scheduler.SparkListenerApplicationEnd
-
- SparkListenerApplicationStart - org.apache.spark.scheduler中的类
-
- SparkListenerApplicationStart(String, Option<String>, long, String, Option<String>, Option<Map<String, String>>) - 类 的构造器org.apache.spark.scheduler.SparkListenerApplicationStart
-
- SparkListenerBlockManagerAdded - org.apache.spark.scheduler中的类
-
- SparkListenerBlockManagerAdded(long, BlockManagerId, long, Option<Object>, Option<Object>) - 类 的构造器org.apache.spark.scheduler.SparkListenerBlockManagerAdded
-
- SparkListenerBlockManagerRemoved - org.apache.spark.scheduler中的类
-
- SparkListenerBlockManagerRemoved(long, BlockManagerId) - 类 的构造器org.apache.spark.scheduler.SparkListenerBlockManagerRemoved
-
- SparkListenerBlockUpdated - org.apache.spark.scheduler中的类
-
- SparkListenerBlockUpdated(BlockUpdatedInfo) - 类 的构造器org.apache.spark.scheduler.SparkListenerBlockUpdated
-
- SparkListenerEnvironmentUpdate - org.apache.spark.scheduler中的类
-
- SparkListenerEnvironmentUpdate(Map<String, Seq<Tuple2<String, String>>>) - 类 的构造器org.apache.spark.scheduler.SparkListenerEnvironmentUpdate
-
- SparkListenerEvent - org.apache.spark.scheduler中的接口
-
- SparkListenerExecutorAdded - org.apache.spark.scheduler中的类
-
- SparkListenerExecutorAdded(long, String, ExecutorInfo) - 类 的构造器org.apache.spark.scheduler.SparkListenerExecutorAdded
-
- SparkListenerExecutorBlacklisted - org.apache.spark.scheduler中的类
-
- SparkListenerExecutorBlacklisted(long, String, int) - 类 的构造器org.apache.spark.scheduler.SparkListenerExecutorBlacklisted
-
- SparkListenerExecutorMetricsUpdate - org.apache.spark.scheduler中的类
-
Periodic updates from executors.
- SparkListenerExecutorMetricsUpdate(String, Seq<Tuple4<Object, Object, Object, Seq<AccumulableInfo>>>) - 类 的构造器org.apache.spark.scheduler.SparkListenerExecutorMetricsUpdate
-
- SparkListenerExecutorRemoved - org.apache.spark.scheduler中的类
-
- SparkListenerExecutorRemoved(long, String, String) - 类 的构造器org.apache.spark.scheduler.SparkListenerExecutorRemoved
-
- SparkListenerExecutorUnblacklisted - org.apache.spark.scheduler中的类
-
- SparkListenerExecutorUnblacklisted(long, String) - 类 的构造器org.apache.spark.scheduler.SparkListenerExecutorUnblacklisted
-
- SparkListenerJobEnd - org.apache.spark.scheduler中的类
-
- SparkListenerJobEnd(int, long, JobResult) - 类 的构造器org.apache.spark.scheduler.SparkListenerJobEnd
-
- SparkListenerJobStart - org.apache.spark.scheduler中的类
-
- SparkListenerJobStart(int, long, Seq<StageInfo>, Properties) - 类 的构造器org.apache.spark.scheduler.SparkListenerJobStart
-
- SparkListenerLogStart - org.apache.spark.scheduler中的类
-
An internal class that describes the metadata of an event log.
- SparkListenerLogStart(String) - 类 的构造器org.apache.spark.scheduler.SparkListenerLogStart
-
- SparkListenerNodeBlacklisted - org.apache.spark.scheduler中的类
-
- SparkListenerNodeBlacklisted(long, String, int) - 类 的构造器org.apache.spark.scheduler.SparkListenerNodeBlacklisted
-
- SparkListenerNodeUnblacklisted - org.apache.spark.scheduler中的类
-
- SparkListenerNodeUnblacklisted(long, String) - 类 的构造器org.apache.spark.scheduler.SparkListenerNodeUnblacklisted
-
- SparkListenerSpeculativeTaskSubmitted - org.apache.spark.scheduler中的类
-
- SparkListenerSpeculativeTaskSubmitted(int) - 类 的构造器org.apache.spark.scheduler.SparkListenerSpeculativeTaskSubmitted
-
- SparkListenerStageCompleted - org.apache.spark.scheduler中的类
-
- SparkListenerStageCompleted(StageInfo) - 类 的构造器org.apache.spark.scheduler.SparkListenerStageCompleted
-
- SparkListenerStageSubmitted - org.apache.spark.scheduler中的类
-
- SparkListenerStageSubmitted(StageInfo, Properties) - 类 的构造器org.apache.spark.scheduler.SparkListenerStageSubmitted
-
- SparkListenerTaskEnd - org.apache.spark.scheduler中的类
-
- SparkListenerTaskEnd(int, int, String, TaskEndReason, TaskInfo, TaskMetrics) - 类 的构造器org.apache.spark.scheduler.SparkListenerTaskEnd
-
- SparkListenerTaskGettingResult - org.apache.spark.scheduler中的类
-
- SparkListenerTaskGettingResult(TaskInfo) - 类 的构造器org.apache.spark.scheduler.SparkListenerTaskGettingResult
-
- SparkListenerTaskStart - org.apache.spark.scheduler中的类
-
- SparkListenerTaskStart(int, int, TaskInfo) - 类 的构造器org.apache.spark.scheduler.SparkListenerTaskStart
-
- SparkListenerUnpersistRDD - org.apache.spark.scheduler中的类
-
- SparkListenerUnpersistRDD(int) - 类 的构造器org.apache.spark.scheduler.SparkListenerUnpersistRDD
-
- SparkMasterRegex - org.apache.spark中的类
-
A collection of regexes for extracting information from the master string.
- SparkMasterRegex() - 类 的构造器org.apache.spark.SparkMasterRegex
-
- sparkProperties() - 类 中的方法org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.SparkAppConfig
-
- sparkProperties() - 类 中的方法org.apache.spark.status.api.v1.ApplicationEnvironmentInfo
-
- SparkRDefaults - org.apache.spark.api.r中的类
-
- SparkRDefaults() - 类 的构造器org.apache.spark.api.r.SparkRDefaults
-
- sparkRPackagePath(boolean) - 类 中的静态方法org.apache.spark.api.r.RUtils
-
Get the list of paths for R packages in various deployment modes, of which the first
path is for the SparkR package itself.
- sparkSession() - 接口 中的方法org.apache.spark.ml.classification.BinaryLogisticRegressionSummary
-
- sparkSession() - 类 中的方法org.apache.spark.sql.Dataset
-
- SparkSession - org.apache.spark.sql中的类
-
The entry point to programming Spark with the Dataset and DataFrame API.
- sparkSession() - 类 中的方法org.apache.spark.sql.SQLContext
-
- sparkSession() - 接口 中的方法org.apache.spark.sql.streaming.StreamingQuery
-
Returns the SparkSession associated with this.
- SparkSession.Builder - org.apache.spark.sql中的类
-
- SparkSession.implicits$ - org.apache.spark.sql中的类
-
:: Experimental ::
(Scala-specific) Implicit methods available in Scala for converting
common Scala objects into DataFrames.
- SparkSessionExtensions - org.apache.spark.sql中的类
-
:: Experimental ::
Holder for injection points to the
SparkSession.
- SparkSessionExtensions() - 类 的构造器org.apache.spark.sql.SparkSessionExtensions
-
- SparkShutdownHook - org.apache.spark.util中的类
-
- SparkShutdownHook(int, Function0<BoxedUnit>) - 类 的构造器org.apache.spark.util.SparkShutdownHook
-
- SparkStageInfo - org.apache.spark中的接口
-
Exposes information about Spark Stages.
- SparkStageInfoImpl - org.apache.spark中的类
-
- SparkStageInfoImpl(int, int, long, String, int, int, int, int) - 类 的构造器org.apache.spark.SparkStageInfoImpl
-
- SparkStatusTracker - org.apache.spark中的类
-
Low-level status reporting APIs for monitoring job and stage progress.
- sparkUser() - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
- sparkUser() - 类 中的方法org.apache.spark.scheduler.SparkListenerApplicationStart
-
- sparkUser() - 类 中的方法org.apache.spark.SparkContext
-
- sparkUser() - 类 中的方法org.apache.spark.status.api.v1.ApplicationAttemptInfo
-
- sparkVersion() - 类 中的方法org.apache.spark.scheduler.SparkListenerLogStart
-
- sparse(int, int, int[], int[], double[]) - 类 中的静态方法org.apache.spark.ml.linalg.Matrices
-
Creates a column-major sparse matrix in Compressed Sparse Column (CSC) format.
- sparse(int, int[], double[]) - 类 中的静态方法org.apache.spark.ml.linalg.Vectors
-
Creates a sparse vector providing its index array and value array.
- sparse(int, Seq<Tuple2<Object, Object>>) - 类 中的静态方法org.apache.spark.ml.linalg.Vectors
-
Creates a sparse vector using unordered (index, value) pairs.
- sparse(int, Iterable<Tuple2<Integer, Double>>) - 类 中的静态方法org.apache.spark.ml.linalg.Vectors
-
Creates a sparse vector using unordered (index, value) pairs in a Java friendly way.
- sparse(int, int, int[], int[], double[]) - 类 中的静态方法org.apache.spark.mllib.linalg.Matrices
-
Creates a column-major sparse matrix in Compressed Sparse Column (CSC) format.
- sparse(int, int[], double[]) - 类 中的静态方法org.apache.spark.mllib.linalg.Vectors
-
Creates a sparse vector providing its index array and value array.
- sparse(int, Seq<Tuple2<Object, Object>>) - 类 中的静态方法org.apache.spark.mllib.linalg.Vectors
-
Creates a sparse vector using unordered (index, value) pairs.
- sparse(int, Iterable<Tuple2<Integer, Double>>) - 类 中的静态方法org.apache.spark.mllib.linalg.Vectors
-
Creates a sparse vector using unordered (index, value) pairs in a Java friendly way.
- SparseMatrix - org.apache.spark.ml.linalg中的类
-
Column-major sparse matrix.
- SparseMatrix(int, int, int[], int[], double[], boolean) - 类 的构造器org.apache.spark.ml.linalg.SparseMatrix
-
- SparseMatrix(int, int, int[], int[], double[]) - 类 的构造器org.apache.spark.ml.linalg.SparseMatrix
-
Column-major sparse matrix.
- SparseMatrix - org.apache.spark.mllib.linalg中的类
-
Column-major sparse matrix.
- SparseMatrix(int, int, int[], int[], double[], boolean) - 类 的构造器org.apache.spark.mllib.linalg.SparseMatrix
-
- SparseMatrix(int, int, int[], int[], double[]) - 类 的构造器org.apache.spark.mllib.linalg.SparseMatrix
-
Column-major sparse matrix.
- SparseVector - org.apache.spark.ml.linalg中的类
-
A sparse vector represented by an index array and a value array.
- SparseVector(int, int[], double[]) - 类 的构造器org.apache.spark.ml.linalg.SparseVector
-
- SparseVector - org.apache.spark.mllib.linalg中的类
-
A sparse vector represented by an index array and a value array.
- SparseVector(int, int[], double[]) - 类 的构造器org.apache.spark.mllib.linalg.SparseVector
-
- SPARSITY() - 类 中的静态方法org.apache.spark.ml.attribute.AttributeKeys
-
- sparsity() - 类 中的方法org.apache.spark.ml.attribute.NumericAttribute
-
- spdiag(Vector) - 类 中的静态方法org.apache.spark.ml.linalg.SparseMatrix
-
Generate a diagonal matrix in SparseMatrix format from the supplied values.
- spdiag(Vector) - 类 中的静态方法org.apache.spark.mllib.linalg.SparseMatrix
-
Generate a diagonal matrix in SparseMatrix format from the supplied values.
- SpearmanCorrelation - org.apache.spark.mllib.stat.correlation中的类
-
Compute Spearman's correlation for two RDDs of the type RDD[Double] or the correlation matrix
for an RDD of the type RDD[Vector].
- SpearmanCorrelation() - 类 的构造器org.apache.spark.mllib.stat.correlation.SpearmanCorrelation
-
- SpecialLengths - org.apache.spark.api.r中的类
-
- SpecialLengths() - 类 的构造器org.apache.spark.api.r.SpecialLengths
-
- speculative() - 类 中的方法org.apache.spark.scheduler.TaskInfo
-
- speculative() - 类 中的方法org.apache.spark.status.api.v1.TaskData
-
- speye(int) - 类 中的静态方法org.apache.spark.ml.linalg.Matrices
-
Generate a sparse Identity Matrix in Matrix format.
- speye(int) - 类 中的静态方法org.apache.spark.ml.linalg.SparseMatrix
-
Generate an Identity Matrix in SparseMatrix format.
- speye(int) - 类 中的静态方法org.apache.spark.mllib.linalg.Matrices
-
Generate a sparse Identity Matrix in Matrix format.
- speye(int) - 类 中的静态方法org.apache.spark.mllib.linalg.SparseMatrix
-
Generate an Identity Matrix in SparseMatrix format.
- SpillListener - org.apache.spark中的类
-
A SparkListener that detects whether spills have occurred in Spark jobs.
- SpillListener() - 类 的构造器org.apache.spark.SpillListener
-
- split() - 类 中的方法org.apache.spark.ml.tree.DecisionTreeModelReadWrite.NodeData
-
- split() - 类 中的方法org.apache.spark.ml.tree.InternalNode
-
- Split - org.apache.spark.ml.tree中的接口
-
Interface for a "Split," which specifies a test made at a decision tree node
to choose the left or right path.
- split() - 类 中的方法org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.NodeData
-
- split() - 类 中的方法org.apache.spark.mllib.tree.model.Node
-
- Split - org.apache.spark.mllib.tree.model中的类
-
:: DeveloperApi ::
Split applied to a feature
param: feature feature index
param: threshold Threshold for continuous feature.
- Split(int, double, Enumeration.Value, List<Object>) - 类 的构造器org.apache.spark.mllib.tree.model.Split
-
- split(Column, String) - 类 中的静态方法org.apache.spark.sql.functions
-
Splits str around pattern (pattern is a regular expression).
- splitAndCountPartitions(Iterator<String>) - 类 中的静态方法org.apache.spark.streaming.util.RawTextHelper
-
Splits lines and counts the words.
- splitAt(int) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- splitCommandString(String) - 类 中的静态方法org.apache.spark.util.Utils
-
Split a string of potentially quoted arguments from the command line the way that a shell
would do it to determine arguments to a command.
- SplitData(int, double[], int) - 类 的构造器org.apache.spark.ml.tree.DecisionTreeModelReadWrite.SplitData
-
- SplitData(int, double, int, Seq<Object>) - 类 的构造器org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.SplitData
-
- SplitData$() - 类 的构造器org.apache.spark.ml.tree.DecisionTreeModelReadWrite.SplitData$
-
- splitIndex() - 类 中的方法org.apache.spark.storage.RDDBlockId
-
- SplitInfo - org.apache.spark.scheduler中的类
-
- SplitInfo(Class<?>, String, String, long, Object) - 类 的构造器org.apache.spark.scheduler.SplitInfo
-
- splits() - 类 中的方法org.apache.spark.ml.feature.Bucketizer
-
Parameter for mapping continuous features into buckets.
- splitsArray() - 类 中的方法org.apache.spark.ml.feature.Bucketizer
-
Parameter for specifying multiple splits parameters.
- spr(double, Vector, DenseVector) - 类 中的静态方法org.apache.spark.ml.linalg.BLAS
-
Adds alpha * x * x.t to a matrix in-place.
- spr(double, Vector, double[]) - 类 中的静态方法org.apache.spark.ml.linalg.BLAS
-
Adds alpha * x * x.t to a matrix in-place.
- spr(double, Vector, DenseVector) - 类 中的静态方法org.apache.spark.mllib.linalg.BLAS
-
Adds alpha * v * v.t to a matrix in-place.
- spr(double, Vector, double[]) - 类 中的静态方法org.apache.spark.mllib.linalg.BLAS
-
Adds alpha * v * v.t to a matrix in-place.
- sprand(int, int, double, Random) - 类 中的静态方法org.apache.spark.ml.linalg.Matrices
-
Generate a SparseMatrix consisting of i.i.d.
- sprand(int, int, double, Random) - 类 中的静态方法org.apache.spark.ml.linalg.SparseMatrix
-
Generate a SparseMatrix consisting of i.i.d. uniform random numbers.
- sprand(int, int, double, Random) - 类 中的静态方法org.apache.spark.mllib.linalg.Matrices
-
Generate a SparseMatrix consisting of i.i.d.
- sprand(int, int, double, Random) - 类 中的静态方法org.apache.spark.mllib.linalg.SparseMatrix
-
Generate a SparseMatrix consisting of i.i.d. uniform random numbers.
- sprandn(int, int, double, Random) - 类 中的静态方法org.apache.spark.ml.linalg.Matrices
-
Generate a SparseMatrix consisting of i.i.d.
- sprandn(int, int, double, Random) - 类 中的静态方法org.apache.spark.ml.linalg.SparseMatrix
-
Generate a SparseMatrix consisting of i.i.d. gaussian random numbers.
- sprandn(int, int, double, Random) - 类 中的静态方法org.apache.spark.mllib.linalg.Matrices
-
Generate a SparseMatrix consisting of i.i.d.
- sprandn(int, int, double, Random) - 类 中的静态方法org.apache.spark.mllib.linalg.SparseMatrix
-
Generate a SparseMatrix consisting of i.i.d. gaussian random numbers.
- sqdist(Vector, Vector) - 类 中的静态方法org.apache.spark.ml.linalg.Vectors
-
Returns the squared distance between two Vectors.
- sqdist(Vector, Vector) - 类 中的静态方法org.apache.spark.mllib.linalg.Vectors
-
Returns the squared distance between two Vectors.
- sql(String) - 类 中的方法org.apache.spark.sql.SparkSession
-
Executes a SQL query using Spark, returning the result as a DataFrame.
- sql(String) - 类 中的方法org.apache.spark.sql.SQLContext
-
- sql() - 类 中的方法org.apache.spark.sql.types.ArrayType
-
- sql() - 类 中的静态方法org.apache.spark.sql.types.BinaryType
-
- sql() - 类 中的静态方法org.apache.spark.sql.types.BooleanType
-
- sql() - 类 中的静态方法org.apache.spark.sql.types.ByteType
-
- sql() - 类 中的静态方法org.apache.spark.sql.types.CalendarIntervalType
-
- sql() - 类 中的静态方法org.apache.spark.sql.types.CharType
-
- sql() - 类 中的方法org.apache.spark.sql.types.DataType
-
- sql() - 类 中的静态方法org.apache.spark.sql.types.DateType
-
- sql() - 类 中的方法org.apache.spark.sql.types.DecimalType
-
- sql() - 类 中的静态方法org.apache.spark.sql.types.DoubleType
-
- sql() - 类 中的静态方法org.apache.spark.sql.types.FloatType
-
- sql() - 类 中的静态方法org.apache.spark.sql.types.HiveStringType
-
- sql() - 类 中的静态方法org.apache.spark.sql.types.IntegerType
-
- sql() - 类 中的静态方法org.apache.spark.sql.types.LongType
-
- sql() - 类 中的方法org.apache.spark.sql.types.MapType
-
- sql() - 类 中的静态方法org.apache.spark.sql.types.NullType
-
- sql() - 类 中的静态方法org.apache.spark.sql.types.NumericType
-
- sql() - 类 中的静态方法org.apache.spark.sql.types.ObjectType
-
- sql() - 类 中的静态方法org.apache.spark.sql.types.ShortType
-
- sql() - 类 中的静态方法org.apache.spark.sql.types.StringType
-
- sql() - 类 中的方法org.apache.spark.sql.types.StructType
-
- sql() - 类 中的静态方法org.apache.spark.sql.types.TimestampType
-
- sql() - 类 中的静态方法org.apache.spark.sql.types.VarcharType
-
- sqlContext() - 类 中的方法org.apache.spark.sql.Dataset
-
- sqlContext() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- sqlContext() - 类 中的方法org.apache.spark.sql.sources.BaseRelation
-
- sqlContext() - 类 中的方法org.apache.spark.sql.SparkSession
-
A wrapped version of this session in the form of a
SQLContext, for backward compatibility.
- SQLContext - org.apache.spark.sql中的类
-
The entry point for working with structured data (rows and columns) in Spark 1.x.
- SQLContext(SparkContext) - 类 的构造器org.apache.spark.sql.SQLContext
-
- SQLContext(JavaSparkContext) - 类 的构造器org.apache.spark.sql.SQLContext
-
- SQLContext.implicits$ - org.apache.spark.sql中的类
-
:: Experimental ::
(Scala-specific) Implicit methods available in Scala for converting
common Scala objects into DataFrames.
- SQLDataTypes - org.apache.spark.ml.linalg中的类
-
:: DeveloperApi ::
SQL data types for vectors and matrices.
- SQLDataTypes() - 类 的构造器org.apache.spark.ml.linalg.SQLDataTypes
-
- SQLImplicits - org.apache.spark.sql中的类
-
A collection of implicit methods for converting common Scala objects into
Datasets.
- SQLImplicits() - 类 的构造器org.apache.spark.sql.SQLImplicits
-
- SQLImplicits.StringToColumn - org.apache.spark.sql中的类
-
Converts $"col name" into a
Column.
- SQLTransformer - org.apache.spark.ml.feature中的类
-
Implements the transformations which are defined by SQL statement.
- SQLTransformer(String) - 类 的构造器org.apache.spark.ml.feature.SQLTransformer
-
- SQLTransformer() - 类 的构造器org.apache.spark.ml.feature.SQLTransformer
-
- sqlType() - 类 中的方法org.apache.spark.mllib.linalg.VectorUDT
-
- SQLUserDefinedType - org.apache.spark.sql.types中的注释类型
-
::DeveloperApi::
A user-defined type which can be automatically recognized by a SQLContext and registered.
- SQLUtils - org.apache.spark.sql.api.r中的类
-
- SQLUtils() - 类 的构造器org.apache.spark.sql.api.r.SQLUtils
-
- sqrt(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Computes the square root of the specified float value.
- sqrt(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Computes the square root of the specified float value.
- Sqrt$() - 类 的构造器org.apache.spark.ml.regression.GeneralizedLinearRegression.Sqrt$
-
- SquaredError - org.apache.spark.mllib.tree.loss中的类
-
:: DeveloperApi ::
Class for squared error loss calculation.
- SquaredError() - 类 的构造器org.apache.spark.mllib.tree.loss.SquaredError
-
- SquaredEuclideanSilhouette - org.apache.spark.ml.evaluation中的类
-
SquaredEuclideanSilhouette computes the average of the
Silhouette over all the data of the dataset, which is
a measure of how appropriately the data have been clustered.
- SquaredEuclideanSilhouette() - 类 的构造器org.apache.spark.ml.evaluation.SquaredEuclideanSilhouette
-
- SquaredEuclideanSilhouette.ClusterStats - org.apache.spark.ml.evaluation中的类
-
- SquaredEuclideanSilhouette.ClusterStats$ - org.apache.spark.ml.evaluation中的类
-
- SquaredL2Updater - org.apache.spark.mllib.optimization中的类
-
:: DeveloperApi ::
Updater for L2 regularized problems.
- SquaredL2Updater() - 类 的构造器org.apache.spark.mllib.optimization.SquaredL2Updater
-
- squaredNormSum() - 类 中的方法org.apache.spark.ml.evaluation.SquaredEuclideanSilhouette.ClusterStats
-
- Src - 类 中的静态变量org.apache.spark.graphx.TripletFields
-
Expose the source and edge fields but not the destination field.
- srcAttr() - 类 中的方法org.apache.spark.graphx.EdgeContext
-
The vertex attribute of the edge's source vertex.
- srcAttr() - 类 中的方法org.apache.spark.graphx.EdgeTriplet
-
The source vertex attribute
- srcAttr() - 类 中的方法org.apache.spark.graphx.impl.AggregatingEdgeContext
-
- srcId() - 类 中的方法org.apache.spark.graphx.Edge
-
- srcId() - 类 中的方法org.apache.spark.graphx.EdgeContext
-
The vertex id of the edge's source vertex.
- srcId() - 类 中的方法org.apache.spark.graphx.impl.AggregatingEdgeContext
-
- srdd() - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
- ssc() - 类 中的方法org.apache.spark.streaming.api.java.JavaStreamingContext
-
- stackTrace() - 类 中的方法org.apache.spark.ExceptionFailure
-
- stackTraceFromJson(JsonAST.JValue) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- stackTraceToJson(StackTraceElement[]) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- stage() - 类 中的方法org.apache.spark.scheduler.AskPermissionToCommitOutput
-
- STAGE() - 类 中的静态方法org.apache.spark.status.TaskIndexNames
-
- STAGE_DAG() - 类 中的静态方法org.apache.spark.ui.ToolTips
-
- STAGE_TIMELINE() - 类 中的静态方法org.apache.spark.ui.ToolTips
-
- stageAttemptId() - 类 中的方法org.apache.spark.scheduler.SparkListenerTaskEnd
-
- stageAttemptId() - 类 中的方法org.apache.spark.scheduler.SparkListenerTaskStart
-
- stageAttemptNumber() - 类 中的方法org.apache.spark.TaskContext
-
How many times the stage that this task belongs to has been attempted.
- stageCompletedFromJson(JsonAST.JValue) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- stageCompletedToJson(SparkListenerStageCompleted) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- StageData - org.apache.spark.status.api.v1中的类
-
- stageFailed(String) - 类 中的方法org.apache.spark.scheduler.StageInfo
-
- stageId() - 类 中的方法org.apache.spark.scheduler.SparkListenerSpeculativeTaskSubmitted
-
- stageId() - 类 中的方法org.apache.spark.scheduler.SparkListenerTaskEnd
-
- stageId() - 类 中的方法org.apache.spark.scheduler.SparkListenerTaskStart
-
- stageId() - 类 中的方法org.apache.spark.scheduler.StageInfo
-
- stageId() - 接口 中的方法org.apache.spark.SparkStageInfo
-
- stageId() - 类 中的方法org.apache.spark.SparkStageInfoImpl
-
- stageId() - 类 中的方法org.apache.spark.status.api.v1.StageData
-
- stageId() - 类 中的方法org.apache.spark.TaskContext
-
The ID of the stage that this task belong to.
- stageIds() - 类 中的方法org.apache.spark.scheduler.SparkListenerJobStart
-
- stageIds() - 接口 中的方法org.apache.spark.SparkJobInfo
-
- stageIds() - 类 中的方法org.apache.spark.SparkJobInfoImpl
-
- stageIds() - 类 中的方法org.apache.spark.status.api.v1.JobData
-
- stageIds() - 类 中的方法org.apache.spark.status.LiveJob
-
- stageIds() - 类 中的方法org.apache.spark.status.SchedulerPool
-
- stageInfo() - 类 中的方法org.apache.spark.scheduler.SparkListenerStageCompleted
-
- stageInfo() - 类 中的方法org.apache.spark.scheduler.SparkListenerStageSubmitted
-
- StageInfo - org.apache.spark.scheduler中的类
-
:: DeveloperApi ::
Stores information about a stage to pass from the scheduler to SparkListeners.
- StageInfo(int, int, String, int, Seq<RDDInfo>, Seq<Object>, String, TaskMetrics, Seq<Seq<TaskLocation>>) - 类 的构造器org.apache.spark.scheduler.StageInfo
-
- stageInfoFromJson(JsonAST.JValue) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
--------------------------------------------------------------------- *
JSON deserialization methods for classes SparkListenerEvents depend on |
- stageInfos() - 类 中的方法org.apache.spark.scheduler.SparkListenerJobStart
-
- stageInfoToJson(StageInfo) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
------------------------------------------------------------------- *
JSON serialization methods for classes SparkListenerEvents depend on |
- stages() - 类 中的方法org.apache.spark.ml.Pipeline
-
param for pipeline stages
- stages() - 类 中的方法org.apache.spark.ml.PipelineModel
-
- StageStatus - org.apache.spark.status.api.v1中的枚举
-
- stageSubmittedFromJson(JsonAST.JValue) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- stageSubmittedToJson(SparkListenerStageSubmitted) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- standardization() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- standardization() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- standardization() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- standardization() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- standardization() - 接口 中的方法org.apache.spark.ml.param.shared.HasStandardization
-
Param for whether to standardize the training features before fitting the model.
- standardization() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- standardization() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- StandardNormalGenerator - org.apache.spark.mllib.random中的类
-
:: DeveloperApi ::
Generates i.i.d. samples from the standard normal distribution.
- StandardNormalGenerator() - 类 的构造器org.apache.spark.mllib.random.StandardNormalGenerator
-
- StandardScaler - org.apache.spark.ml.feature中的类
-
Standardizes features by removing the mean and scaling to unit variance using column summary
statistics on the samples in the training set.
- StandardScaler(String) - 类 的构造器org.apache.spark.ml.feature.StandardScaler
-
- StandardScaler() - 类 的构造器org.apache.spark.ml.feature.StandardScaler
-
- StandardScaler - org.apache.spark.mllib.feature中的类
-
Standardizes features by removing the mean and scaling to unit std using column summary
statistics on the samples in the training set.
- StandardScaler(boolean, boolean) - 类 的构造器org.apache.spark.mllib.feature.StandardScaler
-
- StandardScaler() - 类 的构造器org.apache.spark.mllib.feature.StandardScaler
-
- StandardScalerModel - org.apache.spark.ml.feature中的类
-
- StandardScalerModel - org.apache.spark.mllib.feature中的类
-
Represents a StandardScaler model that can transform vectors.
- StandardScalerModel(Vector, Vector, boolean, boolean) - 类 的构造器org.apache.spark.mllib.feature.StandardScalerModel
-
- StandardScalerModel(Vector, Vector) - 类 的构造器org.apache.spark.mllib.feature.StandardScalerModel
-
- StandardScalerModel(Vector) - 类 的构造器org.apache.spark.mllib.feature.StandardScalerModel
-
- starGraph(SparkContext, int) - 类 中的静态方法org.apache.spark.graphx.util.GraphGenerators
-
Create a star graph with vertex 0 being the center.
- start(String) - 类 中的方法org.apache.spark.sql.streaming.DataStreamWriter
-
Starts the execution of the streaming query, which will continually output results to the given
path as new data arrives.
- start() - 类 中的方法org.apache.spark.sql.streaming.DataStreamWriter
-
Starts the execution of the streaming query, which will continually output results to the given
path as new data arrives.
- start() - 类 中的方法org.apache.spark.streaming.api.java.JavaStreamingContext
-
Start the execution of the streams.
- start() - 类 中的方法org.apache.spark.streaming.dstream.ConstantInputDStream
-
- start() - 类 中的方法org.apache.spark.streaming.dstream.InputDStream
-
Method called to start receiving data.
- start() - 类 中的方法org.apache.spark.streaming.dstream.ReceiverInputDStream
-
- start() - 类 中的方法org.apache.spark.streaming.StreamingContext
-
Start the execution of the streams.
- startApplication(SparkAppHandle.Listener...) - 类 中的方法org.apache.spark.launcher.AbstractLauncher
-
Starts a Spark application.
- startApplication(SparkAppHandle.Listener...) - 类 中的方法org.apache.spark.launcher.InProcessLauncher
-
Starts a Spark application.
- startApplication(SparkAppHandle.Listener...) - 类 中的方法org.apache.spark.launcher.SparkLauncher
-
Starts a Spark application.
- startIndexInLevel(int) - 类 中的静态方法org.apache.spark.mllib.tree.model.Node
-
Return the index of the first node in the given level.
- startJettyServer(String, int, org.apache.spark.SSLOptions, Seq<ServletContextHandler>, SparkConf, String) - 类 中的静态方法org.apache.spark.ui.JettyUtils
-
Attempt to start a Jetty server bound to the supplied hostName:port using the given
context handlers.
- startOffset() - 类 中的方法org.apache.spark.sql.streaming.SourceProgress
-
- startOffset() - 异常错误 中的方法org.apache.spark.sql.streaming.StreamingQueryException
-
- startPosition() - 异常错误 中的方法org.apache.spark.sql.AnalysisException
-
- startServiceOnPort(int, Function1<Object, Tuple2<T, Object>>, SparkConf, String) - 类 中的静态方法org.apache.spark.util.Utils
-
Attempt to start a service on the given port, or fail after a number of attempts.
- startsWith(Column) - 类 中的方法org.apache.spark.sql.Column
-
String starts with.
- startsWith(String) - 类 中的方法org.apache.spark.sql.Column
-
String starts with another string literal.
- startsWith(GenSeq<B>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- startsWith(GenSeq<B>, int) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- startTime() - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
- startTime() - 类 中的方法org.apache.spark.SparkContext
-
- startTime() - 类 中的方法org.apache.spark.status.api.v1.ApplicationAttemptInfo
-
- startTime() - 类 中的方法org.apache.spark.status.api.v1.streaming.OutputOperationInfo
-
- startTime() - 类 中的方法org.apache.spark.status.api.v1.streaming.StreamingStatistics
-
- startTime() - 类 中的方法org.apache.spark.streaming.scheduler.OutputOperationInfo
-
- stat() - 类 中的方法org.apache.spark.sql.Dataset
-
- StatCounter - org.apache.spark.util中的类
-
A class for tracking the statistics of a set of numbers (count, mean and variance) in a
numerically robust way.
- StatCounter(TraversableOnce<Object>) - 类 的构造器org.apache.spark.util.StatCounter
-
- StatCounter() - 类 的构造器org.apache.spark.util.StatCounter
-
Initialize the StatCounter with no values.
- state() - 类 中的方法org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.StatusUpdate
-
- state() - 类 中的方法org.apache.spark.scheduler.local.StatusUpdate
-
- State<S> - org.apache.spark.streaming中的类
-
:: Experimental ::
Abstract class for getting and updating the state in mapping function used in the
mapWithState
operation of a
pair DStream (Scala)
or a
JavaPairDStream (Java).
- State() - 类 的构造器org.apache.spark.streaming.State
-
- stateChanged(SparkAppHandle) - 接口 中的方法org.apache.spark.launcher.SparkAppHandle.Listener
-
Callback for changes in the handle's state.
- statement() - 类 中的方法org.apache.spark.ml.feature.SQLTransformer
-
SQL statement parameter.
- StateOperatorProgress - org.apache.spark.sql.streaming中的类
-
Information about updates made to stateful operators in a
StreamingQuery during a trigger.
- stateOperators() - 类 中的方法org.apache.spark.sql.streaming.StreamingQueryProgress
-
- stateSnapshots() - 类 中的方法org.apache.spark.streaming.api.java.JavaMapWithStateDStream
-
- stateSnapshots() - 类 中的方法org.apache.spark.streaming.dstream.MapWithStateDStream
-
Return a pair DStream where each RDD is the snapshot of the state of all the keys.
- StateSpec<KeyType,ValueType,StateType,MappedType> - org.apache.spark.streaming中的类
-
:: Experimental ::
Abstract class representing all the specifications of the DStream transformation
mapWithState operation of a
pair DStream (Scala) or a
JavaPairDStream (Java).
- StateSpec() - 类 的构造器org.apache.spark.streaming.StateSpec
-
- staticPageRank(int, double) - 类 中的方法org.apache.spark.graphx.GraphOps
-
Run PageRank for a fixed number of iterations returning a graph with vertex attributes
containing the PageRank and edge attributes the normalized edge weight.
- staticParallelPersonalizedPageRank(long[], int, double) - 类 中的方法org.apache.spark.graphx.GraphOps
-
Run parallel personalized PageRank for a given array of source vertices, such
that all random walks are started relative to the source vertices
- staticPersonalizedPageRank(long, int, double) - 类 中的方法org.apache.spark.graphx.GraphOps
-
Run Personalized PageRank for a fixed number of iterations with
with all iterations originating at the source node
returning a graph with vertex attributes
containing the PageRank and edge attributes the normalized edge weight.
- StaticSources - org.apache.spark.metrics.source中的类
-
- StaticSources() - 类 的构造器org.apache.spark.metrics.source.StaticSources
-
- statistic() - 类 中的方法org.apache.spark.mllib.stat.test.ChiSqTestResult
-
- statistic() - 类 中的方法org.apache.spark.mllib.stat.test.KolmogorovSmirnovTestResult
-
- statistic() - 接口 中的方法org.apache.spark.mllib.stat.test.TestResult
-
Test statistic.
- Statistics - org.apache.spark.mllib.stat中的类
-
API for statistical functions in MLlib.
- Statistics() - 类 的构造器org.apache.spark.mllib.stat.Statistics
-
- Statistics - org.apache.spark.sql.sources.v2.reader中的接口
-
- stats() - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
Return a
StatCounter object that captures the mean, variance and
count of the RDD's elements in one operation.
- stats() - 类 中的方法org.apache.spark.mllib.tree.model.Node
-
- stats() - 类 中的方法org.apache.spark.rdd.DoubleRDDFunctions
-
Return a
StatCounter object that captures the mean, variance and
count of the RDD's elements in one operation.
- stats() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- stats() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- stats() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- StatsdMetricType - org.apache.spark.metrics.sink中的类
-
- StatsdMetricType() - 类 的构造器org.apache.spark.metrics.sink.StatsdMetricType
-
- StatsReportListener - org.apache.spark.scheduler中的类
-
:: DeveloperApi ::
Simple SparkListener that logs a few summary statistics when each stage completes.
- StatsReportListener() - 类 的构造器org.apache.spark.scheduler.StatsReportListener
-
- StatsReportListener - org.apache.spark.streaming.scheduler中的类
-
:: DeveloperApi ::
A simple StreamingListener that logs summary statistics across Spark Streaming batches
param: numBatchInfos Number of last batches to consider for generating statistics (default: 10)
- StatsReportListener(int) - 类 的构造器org.apache.spark.streaming.scheduler.StatsReportListener
-
- status() - 类 中的方法org.apache.spark.scheduler.TaskInfo
-
- status() - 接口 中的方法org.apache.spark.SparkJobInfo
-
- status() - 类 中的方法org.apache.spark.SparkJobInfoImpl
-
- status() - 接口 中的方法org.apache.spark.sql.streaming.StreamingQuery
-
Returns the current status of the query.
- status() - 类 中的方法org.apache.spark.status.api.v1.JobData
-
- status() - 类 中的方法org.apache.spark.status.api.v1.StageData
-
- status() - 类 中的方法org.apache.spark.status.api.v1.streaming.BatchInfo
-
- status() - 类 中的方法org.apache.spark.status.api.v1.TaskData
-
- status() - 类 中的方法org.apache.spark.status.LiveJob
-
- status() - 类 中的方法org.apache.spark.status.LiveStage
-
- STATUS() - 类 中的静态方法org.apache.spark.status.TaskIndexNames
-
- status() - 类 中的方法org.apache.spark.storage.BlockManagerMessages.BlockLocationsAndStatus
-
- statusTracker() - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
- statusTracker() - 类 中的方法org.apache.spark.SparkContext
-
- StatusUpdate(String, long, Enumeration.Value, org.apache.spark.util.SerializableBuffer) - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.StatusUpdate
-
- StatusUpdate - org.apache.spark.scheduler.local中的类
-
- StatusUpdate(long, Enumeration.Value, ByteBuffer) - 类 的构造器org.apache.spark.scheduler.local.StatusUpdate
-
- StatusUpdate$() - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.StatusUpdate$
-
- STD() - 类 中的静态方法org.apache.spark.ml.attribute.AttributeKeys
-
- std() - 类 中的方法org.apache.spark.ml.attribute.NumericAttribute
-
- std() - 类 中的方法org.apache.spark.ml.feature.StandardScalerModel
-
- std() - 类 中的方法org.apache.spark.mllib.feature.StandardScalerModel
-
- std() - 类 中的方法org.apache.spark.mllib.random.LogNormalGenerator
-
- stddev(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: alias for stddev_samp.
- stddev(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: alias for stddev_samp.
- stddev_pop(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the population standard deviation of
the expression in a group.
- stddev_pop(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the population standard deviation of
the expression in a group.
- stddev_samp(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the sample standard deviation of
the expression in a group.
- stddev_samp(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the sample standard deviation of
the expression in a group.
- stdev() - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
Compute the population standard deviation of this RDD's elements.
- stdev() - 类 中的方法org.apache.spark.rdd.DoubleRDDFunctions
-
Compute the population standard deviation of this RDD's elements.
- stdev() - 类 中的方法org.apache.spark.util.StatCounter
-
Return the population standard deviation of the values.
- stepSize() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- stepSize() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- stepSize() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- stepSize() - 类 中的静态方法org.apache.spark.ml.feature.Word2Vec
-
- stepSize() - 类 中的静态方法org.apache.spark.ml.feature.Word2VecModel
-
- stepSize() - 接口 中的方法org.apache.spark.ml.param.shared.HasStepSize
-
Param for Step size to be used for each iteration of optimization (> 0).
- stepSize() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- stepSize() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- stop() - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
Shut down the SparkContext.
- stop() - 接口 中的方法org.apache.spark.launcher.SparkAppHandle
-
Asks the application to stop.
- stop() - 类 中的方法org.apache.spark.SparkContext
-
Shut down the SparkContext.
- stop() - 类 中的方法org.apache.spark.sql.SparkSession
-
Stop the underlying SparkContext.
- stop() - 接口 中的方法org.apache.spark.sql.streaming.StreamingQuery
-
Stops the execution of this query if it is running.
- stop() - 类 中的方法org.apache.spark.streaming.api.java.JavaStreamingContext
-
Stop the execution of the streams.
- stop(boolean) - 类 中的方法org.apache.spark.streaming.api.java.JavaStreamingContext
-
Stop the execution of the streams.
- stop(boolean, boolean) - 类 中的方法org.apache.spark.streaming.api.java.JavaStreamingContext
-
Stop the execution of the streams.
- stop() - 类 中的方法org.apache.spark.streaming.dstream.ConstantInputDStream
-
- stop() - 类 中的方法org.apache.spark.streaming.dstream.InputDStream
-
Method called to stop receiving data.
- stop() - 类 中的方法org.apache.spark.streaming.dstream.ReceiverInputDStream
-
- stop(String) - 类 中的方法org.apache.spark.streaming.receiver.Receiver
-
Stop the receiver completely.
- stop(String, Throwable) - 类 中的方法org.apache.spark.streaming.receiver.Receiver
-
Stop the receiver completely due to an exception
- stop(boolean) - 类 中的方法org.apache.spark.streaming.StreamingContext
-
Stop the execution of the streams immediately (does not wait for all received data
to be processed).
- stop(boolean, boolean) - 类 中的方法org.apache.spark.streaming.StreamingContext
-
Stop the execution of the streams, with option of ensuring all received data
has been processed.
- StopAllReceivers - org.apache.spark.streaming.scheduler中的类
-
This message will trigger ReceiverTrackerEndpoint to send stop signals to all registered
receivers.
- StopAllReceivers() - 类 的构造器org.apache.spark.streaming.scheduler.StopAllReceivers
-
- StopBlockManagerMaster$() - 类 的构造器org.apache.spark.storage.BlockManagerMessages.StopBlockManagerMaster$
-
- StopCoordinator - org.apache.spark.scheduler中的类
-
- StopCoordinator() - 类 的构造器org.apache.spark.scheduler.StopCoordinator
-
- StopDriver$() - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.StopDriver$
-
- StopExecutor - org.apache.spark.scheduler.local中的类
-
- StopExecutor() - 类 的构造器org.apache.spark.scheduler.local.StopExecutor
-
- StopExecutor$() - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.StopExecutor$
-
- StopExecutors$() - 类 的构造器org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.StopExecutors$
-
- StopMapOutputTracker - org.apache.spark中的类
-
- StopMapOutputTracker() - 类 的构造器org.apache.spark.StopMapOutputTracker
-
- StopReceiver - org.apache.spark.streaming.receiver中的类
-
- StopReceiver() - 类 的构造器org.apache.spark.streaming.receiver.StopReceiver
-
- stopWords() - 类 中的方法org.apache.spark.ml.feature.StopWordsRemover
-
The words to be filtered out.
- StopWordsRemover - org.apache.spark.ml.feature中的类
-
A feature transformer that filters out stop words from input.
- StopWordsRemover(String) - 类 的构造器org.apache.spark.ml.feature.StopWordsRemover
-
- StopWordsRemover() - 类 的构造器org.apache.spark.ml.feature.StopWordsRemover
-
- storage() - 类 中的方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- STORAGE_MEMORY() - 类 中的静态方法org.apache.spark.ui.ToolTips
-
- storageLevel() - 类 中的方法org.apache.spark.sql.Dataset
-
Get the Dataset's current storage level, or StorageLevel.NONE if not persisted.
- storageLevel() - 类 中的方法org.apache.spark.status.api.v1.RDDPartitionInfo
-
- storageLevel() - 类 中的方法org.apache.spark.status.api.v1.RDDStorageInfo
-
- storageLevel() - 类 中的方法org.apache.spark.status.LiveRDD
-
- storageLevel() - 类 中的方法org.apache.spark.storage.BlockManagerMessages.UpdateBlockInfo
-
- storageLevel() - 类 中的方法org.apache.spark.storage.BlockStatus
-
- storageLevel() - 类 中的方法org.apache.spark.storage.BlockUpdatedInfo
-
- storageLevel() - 类 中的方法org.apache.spark.storage.RDDInfo
-
- StorageLevel - org.apache.spark.storage中的类
-
:: DeveloperApi ::
Flags for controlling the storage of an RDD.
- StorageLevel() - 类 的构造器org.apache.spark.storage.StorageLevel
-
- storageLevel() - 类 中的方法org.apache.spark.streaming.receiver.Receiver
-
- storageLevelFromJson(JsonAST.JValue) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- StorageLevels - org.apache.spark.api.java中的类
-
Expose some commonly useful storage level constants.
- StorageLevels() - 类 的构造器org.apache.spark.api.java.StorageLevels
-
- storageLevelToJson(StorageLevel) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- StorageStatus - org.apache.spark.storage中的类
-
- StorageStatus(BlockManagerId, long, Option<Object>, Option<Object>) - 类 的构造器org.apache.spark.storage.StorageStatus
-
已过时。
- StorageStatus(BlockManagerId, long, Option<Object>, Option<Object>, Map<BlockId, BlockStatus>) - 类 的构造器org.apache.spark.storage.StorageStatus
-
已过时。
Create a storage status with an initial set of blocks, leaving the source unmodified.
- StorageUtils - org.apache.spark.storage中的类
-
Helper methods for storage-related objects.
- StorageUtils() - 类 的构造器org.apache.spark.storage.StorageUtils
-
- store(T) - 类 中的方法org.apache.spark.streaming.receiver.Receiver
-
Store a single item of received data to Spark's memory.
- store(ArrayBuffer<T>) - 类 中的方法org.apache.spark.streaming.receiver.Receiver
-
Store an ArrayBuffer of received data as a data block into Spark's memory.
- store(ArrayBuffer<T>, Object) - 类 中的方法org.apache.spark.streaming.receiver.Receiver
-
Store an ArrayBuffer of received data as a data block into Spark's memory.
- store(Iterator<T>) - 类 中的方法org.apache.spark.streaming.receiver.Receiver
-
Store an iterator of received data as a data block into Spark's memory.
- store(Iterator<T>, Object) - 类 中的方法org.apache.spark.streaming.receiver.Receiver
-
Store an iterator of received data as a data block into Spark's memory.
- store(Iterator<T>) - 类 中的方法org.apache.spark.streaming.receiver.Receiver
-
Store an iterator of received data as a data block into Spark's memory.
- store(Iterator<T>, Object) - 类 中的方法org.apache.spark.streaming.receiver.Receiver
-
Store an iterator of received data as a data block into Spark's memory.
- store(ByteBuffer) - 类 中的方法org.apache.spark.streaming.receiver.Receiver
-
Store the bytes of received data as a data block into Spark's memory.
- store(ByteBuffer, Object) - 类 中的方法org.apache.spark.streaming.receiver.Receiver
-
Store the bytes of received data as a data block into Spark's memory.
- strategy() - 类 中的静态方法org.apache.spark.ml.feature.Imputer
-
- strategy() - 类 中的静态方法org.apache.spark.ml.feature.ImputerModel
-
- Strategy - org.apache.spark.mllib.tree.configuration中的类
-
Stores all the configuration options for tree construction
param: algo Learning goal.
- Strategy(Enumeration.Value, Impurity, int, int, int, Enumeration.Value, Map<Object, Object>, int, double, int, double, boolean, int) - 类 的构造器org.apache.spark.mllib.tree.configuration.Strategy
-
- Strategy(Enumeration.Value, Impurity, int, int, int, Map<Integer, Integer>) - 类 的构造器org.apache.spark.mllib.tree.configuration.Strategy
-
- StratifiedSamplingUtils - org.apache.spark.util.random中的类
-
Auxiliary functions and data structures for the sampleByKey method in PairRDDFunctions.
- StratifiedSamplingUtils() - 类 的构造器org.apache.spark.util.random.StratifiedSamplingUtils
-
- STREAM() - 类 中的静态方法org.apache.spark.storage.BlockId
-
- StreamBlockId - org.apache.spark.storage中的类
-
- StreamBlockId(int, long) - 类 的构造器org.apache.spark.storage.StreamBlockId
-
- streamId() - 类 中的方法org.apache.spark.status.api.v1.streaming.ReceiverInfo
-
- streamId() - 类 中的方法org.apache.spark.storage.StreamBlockId
-
- streamId() - 类 中的方法org.apache.spark.streaming.receiver.Receiver
-
Get the unique identifier the receiver input stream that this
receiver is associated with.
- streamId() - 类 中的方法org.apache.spark.streaming.scheduler.ReceiverInfo
-
- streamIdToInputInfo() - 类 中的方法org.apache.spark.streaming.scheduler.BatchInfo
-
- StreamingContext - org.apache.spark.streaming中的类
-
Main entry point for Spark Streaming functionality.
- StreamingContext(SparkContext, Duration) - 类 的构造器org.apache.spark.streaming.StreamingContext
-
Create a StreamingContext using an existing SparkContext.
- StreamingContext(SparkConf, Duration) - 类 的构造器org.apache.spark.streaming.StreamingContext
-
Create a StreamingContext by providing the configuration necessary for a new SparkContext.
- StreamingContext(String, String, Duration, String, Seq<String>, Map<String, String>) - 类 的构造器org.apache.spark.streaming.StreamingContext
-
Create a StreamingContext by providing the details necessary for creating a new SparkContext.
- StreamingContext(String, Configuration) - 类 的构造器org.apache.spark.streaming.StreamingContext
-
Recreate a StreamingContext from a checkpoint file.
- StreamingContext(String) - 类 的构造器org.apache.spark.streaming.StreamingContext
-
Recreate a StreamingContext from a checkpoint file.
- StreamingContext(String, SparkContext) - 类 的构造器org.apache.spark.streaming.StreamingContext
-
Recreate a StreamingContext from a checkpoint file using an existing SparkContext.
- StreamingContextPythonHelper - org.apache.spark.streaming中的类
-
- StreamingContextPythonHelper() - 类 的构造器org.apache.spark.streaming.StreamingContextPythonHelper
-
- StreamingContextState - org.apache.spark.streaming中的枚举
-
:: DeveloperApi ::
Represents the state of a StreamingContext.
- StreamingKMeans - org.apache.spark.mllib.clustering中的类
-
StreamingKMeans provides methods for configuring a
streaming k-means analysis, training the model on streaming,
and using the model to make predictions on streaming data.
- StreamingKMeans(int, double, String) - 类 的构造器org.apache.spark.mllib.clustering.StreamingKMeans
-
- StreamingKMeans() - 类 的构造器org.apache.spark.mllib.clustering.StreamingKMeans
-
- StreamingKMeansModel - org.apache.spark.mllib.clustering中的类
-
StreamingKMeansModel extends MLlib's KMeansModel for streaming
algorithms, so it can keep track of a continuously updated weight
associated with each cluster, and also update the model by
doing a single iteration of the standard k-means algorithm.
- StreamingKMeansModel(Vector[], double[]) - 类 的构造器org.apache.spark.mllib.clustering.StreamingKMeansModel
-
- StreamingLinearAlgorithm<M extends GeneralizedLinearModel,A extends GeneralizedLinearAlgorithm<M>> - org.apache.spark.mllib.regression中的类
-
:: DeveloperApi ::
StreamingLinearAlgorithm implements methods for continuously
training a generalized linear model on streaming data,
and using it for prediction on (possibly different) streaming data.
- StreamingLinearAlgorithm() - 类 的构造器org.apache.spark.mllib.regression.StreamingLinearAlgorithm
-
- StreamingLinearRegressionWithSGD - org.apache.spark.mllib.regression中的类
-
Train or predict a linear regression model on streaming data.
- StreamingLinearRegressionWithSGD() - 类 的构造器org.apache.spark.mllib.regression.StreamingLinearRegressionWithSGD
-
Construct a StreamingLinearRegression object with default parameters:
{stepSize: 0.1, numIterations: 50, miniBatchFraction: 1.0}.
- StreamingListener - org.apache.spark.streaming.scheduler中的接口
-
:: DeveloperApi ::
A listener interface for receiving information about an ongoing streaming
computation.
- StreamingListenerBatchCompleted - org.apache.spark.streaming.scheduler中的类
-
- StreamingListenerBatchCompleted(BatchInfo) - 类 的构造器org.apache.spark.streaming.scheduler.StreamingListenerBatchCompleted
-
- StreamingListenerBatchStarted - org.apache.spark.streaming.scheduler中的类
-
- StreamingListenerBatchStarted(BatchInfo) - 类 的构造器org.apache.spark.streaming.scheduler.StreamingListenerBatchStarted
-
- StreamingListenerBatchSubmitted - org.apache.spark.streaming.scheduler中的类
-
- StreamingListenerBatchSubmitted(BatchInfo) - 类 的构造器org.apache.spark.streaming.scheduler.StreamingListenerBatchSubmitted
-
- StreamingListenerEvent - org.apache.spark.streaming.scheduler中的接口
-
:: DeveloperApi ::
Base trait for events related to StreamingListener
- StreamingListenerOutputOperationCompleted - org.apache.spark.streaming.scheduler中的类
-
- StreamingListenerOutputOperationCompleted(OutputOperationInfo) - 类 的构造器org.apache.spark.streaming.scheduler.StreamingListenerOutputOperationCompleted
-
- StreamingListenerOutputOperationStarted - org.apache.spark.streaming.scheduler中的类
-
- StreamingListenerOutputOperationStarted(OutputOperationInfo) - 类 的构造器org.apache.spark.streaming.scheduler.StreamingListenerOutputOperationStarted
-
- StreamingListenerReceiverError - org.apache.spark.streaming.scheduler中的类
-
- StreamingListenerReceiverError(ReceiverInfo) - 类 的构造器org.apache.spark.streaming.scheduler.StreamingListenerReceiverError
-
- StreamingListenerReceiverStarted - org.apache.spark.streaming.scheduler中的类
-
- StreamingListenerReceiverStarted(ReceiverInfo) - 类 的构造器org.apache.spark.streaming.scheduler.StreamingListenerReceiverStarted
-
- StreamingListenerReceiverStopped - org.apache.spark.streaming.scheduler中的类
-
- StreamingListenerReceiverStopped(ReceiverInfo) - 类 的构造器org.apache.spark.streaming.scheduler.StreamingListenerReceiverStopped
-
- StreamingListenerStreamingStarted - org.apache.spark.streaming.scheduler中的类
-
- StreamingListenerStreamingStarted(long) - 类 的构造器org.apache.spark.streaming.scheduler.StreamingListenerStreamingStarted
-
- StreamingLogisticRegressionWithSGD - org.apache.spark.mllib.classification中的类
-
Train or predict a logistic regression model on streaming data.
- StreamingLogisticRegressionWithSGD() - 类 的构造器org.apache.spark.mllib.classification.StreamingLogisticRegressionWithSGD
-
Construct a StreamingLogisticRegression object with default parameters:
{stepSize: 0.1, numIterations: 50, miniBatchFraction: 1.0, regParam: 0.0}.
- StreamingQuery - org.apache.spark.sql.streaming中的接口
-
A handle to a query that is executing continuously in the background as new data arrives.
- StreamingQueryException - org.apache.spark.sql.streaming中的异常错误
-
- StreamingQueryListener - org.apache.spark.sql.streaming中的类
-
- StreamingQueryListener() - 类 的构造器org.apache.spark.sql.streaming.StreamingQueryListener
-
- StreamingQueryListener.Event - org.apache.spark.sql.streaming中的接口
-
- StreamingQueryListener.QueryProgressEvent - org.apache.spark.sql.streaming中的类
-
Event representing any progress updates in a query.
- StreamingQueryListener.QueryStartedEvent - org.apache.spark.sql.streaming中的类
-
Event representing the start of a query
param: id An unique query id that persists across restarts.
- StreamingQueryListener.QueryTerminatedEvent - org.apache.spark.sql.streaming中的类
-
Event representing that termination of a query.
- StreamingQueryManager - org.apache.spark.sql.streaming中的类
-
- StreamingQueryProgress - org.apache.spark.sql.streaming中的类
-
Information about progress made in the execution of a
StreamingQuery during
a trigger.
- StreamingQueryStatus - org.apache.spark.sql.streaming中的类
-
Reports information about the instantaneous status of a streaming query.
- StreamingStatistics - org.apache.spark.status.api.v1.streaming中的类
-
- StreamingTest - org.apache.spark.mllib.stat.test中的类
-
Performs online 2-sample significance testing for a stream of (Boolean, Double) pairs.
- StreamingTest() - 类 的构造器org.apache.spark.mllib.stat.test.StreamingTest
-
- StreamInputInfo - org.apache.spark.streaming.scheduler中的类
-
:: DeveloperApi ::
Track the information of input stream at specified batch time.
- StreamInputInfo(int, long, Map<String, Object>) - 类 的构造器org.apache.spark.streaming.scheduler.StreamInputInfo
-
- streamName() - 类 中的方法org.apache.spark.status.api.v1.streaming.ReceiverInfo
-
- streams() - 类 中的方法org.apache.spark.sql.SparkSession
-
:: Experimental ::
Returns a StreamingQueryManager that allows managing all the
StreamingQuerys active on this.
- streams() - 类 中的方法org.apache.spark.sql.SQLContext
-
- StreamSinkProvider - org.apache.spark.sql.sources中的接口
-
::Experimental::
Implemented by objects that can produce a streaming Sink for a specific format or system.
- StreamSourceProvider - org.apache.spark.sql.sources中的接口
-
::Experimental::
Implemented by objects that can produce a streaming Source for a specific format or system.
- StreamWriter - org.apache.spark.sql.sources.v2.writer.streaming中的接口
-
- StreamWriteSupport - org.apache.spark.sql.sources.v2中的接口
-
- STRING() - 类 中的静态方法org.apache.spark.api.r.SerializationFormats
-
- string() - 类 中的方法org.apache.spark.sql.ColumnName
-
Creates a new StructField of type string.
- STRING() - 类 中的静态方法org.apache.spark.sql.Encoders
-
An encoder for nullable string type.
- StringAccumulatorParam$() - 类 的构造器org.apache.spark.AccumulatorParam.StringAccumulatorParam$
-
已过时。
- StringArrayParam - org.apache.spark.ml.param中的类
-
:: DeveloperApi ::
Specialized version of Param[Array[String} for Java.
- StringArrayParam(Params, String, String, Function1<String[], Object>) - 类 的构造器org.apache.spark.ml.param.StringArrayParam
-
- StringArrayParam(Params, String, String) - 类 的构造器org.apache.spark.ml.param.StringArrayParam
-
- StringContains - org.apache.spark.sql.sources中的类
-
A filter that evaluates to true iff the attribute evaluates to
a string that contains the string value.
- StringContains(String, String) - 类 的构造器org.apache.spark.sql.sources.StringContains
-
- StringEndsWith - org.apache.spark.sql.sources中的类
-
A filter that evaluates to true iff the attribute evaluates to
a string that starts with value.
- StringEndsWith(String, String) - 类 的构造器org.apache.spark.sql.sources.StringEndsWith
-
- StringIndexer - org.apache.spark.ml.feature中的类
-
A label indexer that maps a string column of labels to an ML column of label indices.
- StringIndexer(String) - 类 的构造器org.apache.spark.ml.feature.StringIndexer
-
- StringIndexer() - 类 的构造器org.apache.spark.ml.feature.StringIndexer
-
- StringIndexerModel - org.apache.spark.ml.feature中的类
-
- StringIndexerModel(String, String[]) - 类 的构造器org.apache.spark.ml.feature.StringIndexerModel
-
- StringIndexerModel(String[]) - 类 的构造器org.apache.spark.ml.feature.StringIndexerModel
-
- stringIndexerOrderType() - 类 中的静态方法org.apache.spark.ml.feature.RFormula
-
- stringIndexerOrderType() - 类 中的静态方法org.apache.spark.ml.feature.RFormulaModel
-
- stringOrderType() - 类 中的静态方法org.apache.spark.ml.feature.StringIndexer
-
- stringOrderType() - 类 中的静态方法org.apache.spark.ml.feature.StringIndexerModel
-
- stringPrefix() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- StringRRDD<T> - org.apache.spark.api.r中的类
-
An RDD that stores R objects as Array[String].
- StringRRDD(RDD<T>, byte[], String, byte[], Object[], ClassTag<T>) - 类 的构造器org.apache.spark.api.r.StringRRDD
-
- StringStartsWith - org.apache.spark.sql.sources中的类
-
A filter that evaluates to true iff the attribute evaluates to
a string that starts with value.
- StringStartsWith(String, String) - 类 的构造器org.apache.spark.sql.sources.StringStartsWith
-
- StringToColumn(StringContext) - 类 的构造器org.apache.spark.sql.SQLImplicits.StringToColumn
-
- stringToSeq(String, Function1<String, T>) - 类 中的静态方法org.apache.spark.internal.config.ConfigHelpers
-
- stringToSeq(String) - 类 中的静态方法org.apache.spark.util.Utils
-
- StringType - 类 中的静态变量org.apache.spark.sql.types.DataTypes
-
Gets the StringType object.
- StringType - org.apache.spark.sql.types中的类
-
The data type representing String values.
- stripDirectory(String) - 类 中的静态方法org.apache.spark.util.Utils
-
Strip the directory from a path name
- stripXSS(String) - 类 中的静态方法org.apache.spark.ui.UIUtils
-
Remove suspicious characters of user input to prevent Cross-Site scripting (XSS) attacks
For more information about XSS testing:
https://www.owasp.org/index.php/XSS_Filter_Evasion_Cheat_Sheet and
https://www.owasp.org/index.php/Testing_for_Reflected_Cross_site_scripting_(OTG-INPVAL-001)
- stronglyConnectedComponents(int) - 类 中的方法org.apache.spark.graphx.GraphOps
-
Compute the strongly connected component (SCC) of each vertex and return a graph with the
vertex value containing the lowest vertex id in the SCC containing that vertex.
- StronglyConnectedComponents - org.apache.spark.graphx.lib中的类
-
Strongly connected components algorithm implementation.
- StronglyConnectedComponents() - 类 的构造器org.apache.spark.graphx.lib.StronglyConnectedComponents
-
- struct(Seq<StructField>) - 类 中的方法org.apache.spark.sql.ColumnName
-
Creates a new StructField of type struct.
- struct(StructType) - 类 中的方法org.apache.spark.sql.ColumnName
-
Creates a new StructField of type struct.
- struct(Column...) - 类 中的静态方法org.apache.spark.sql.functions
-
Creates a new struct column.
- struct(String, String...) - 类 中的静态方法org.apache.spark.sql.functions
-
Creates a new struct column that composes multiple input columns.
- struct(Seq<Column>) - 类 中的静态方法org.apache.spark.sql.functions
-
Creates a new struct column.
- struct(String, Seq<String>) - 类 中的静态方法org.apache.spark.sql.functions
-
Creates a new struct column that composes multiple input columns.
- StructField - org.apache.spark.sql.types中的类
-
A field inside a StructType.
- StructField(String, DataType, boolean, Metadata) - 类 的构造器org.apache.spark.sql.types.StructField
-
- StructType - org.apache.spark.sql.types中的类
-
A
StructType object can be constructed by
StructType(fields: Seq[StructField])
For a
StructType object, one or multiple
StructFields can be extracted by names.
- StructType(StructField[]) - 类 的构造器org.apache.spark.sql.types.StructType
-
- StructType() - 类 的构造器org.apache.spark.sql.types.StructType
-
No-arg constructor for kryo.
- StudentTTest - org.apache.spark.mllib.stat.test中的类
-
Performs Students's 2-sample t-test.
- StudentTTest() - 类 的构造器org.apache.spark.mllib.stat.test.StudentTTest
-
- subexpressionEliminationEnabled() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- subgraph(Function1<EdgeTriplet<VD, ED>, Object>, Function2<Object, VD, Object>) - 类 中的方法org.apache.spark.graphx.Graph
-
Restricts the graph to only the vertices and edges satisfying the predicates.
- subgraph(Function1<EdgeTriplet<VD, ED>, Object>, Function2<Object, VD, Object>) - 类 中的方法org.apache.spark.graphx.impl.GraphImpl
-
- subgraph$default$1() - 类 中的静态方法org.apache.spark.graphx.impl.GraphImpl
-
- subgraph$default$2() - 类 中的静态方法org.apache.spark.graphx.impl.GraphImpl
-
- submissionTime() - 类 中的方法org.apache.spark.scheduler.StageInfo
-
When this stage was submitted from the DAGScheduler to a TaskScheduler.
- submissionTime() - 接口 中的方法org.apache.spark.SparkStageInfo
-
- submissionTime() - 类 中的方法org.apache.spark.SparkStageInfoImpl
-
- submissionTime() - 类 中的方法org.apache.spark.status.api.v1.JobData
-
- submissionTime() - 类 中的方法org.apache.spark.status.api.v1.StageData
-
- submissionTime() - 类 中的方法org.apache.spark.streaming.scheduler.BatchInfo
-
- submitJob(RDD<T>, Function1<Iterator<T>, U>, Seq<Object>, Function2<Object, U, BoxedUnit>, Function0<R>) - 接口 中的方法org.apache.spark.JobSubmitter
-
Submit a job for execution and return a FutureAction holding the result.
- submitJob(RDD<T>, Function1<Iterator<T>, U>, Seq<Object>, Function2<Object, U, BoxedUnit>, Function0<R>) - 类 中的方法org.apache.spark.SparkContext
-
Submit a job for execution and return a FutureJob holding the result.
- subModels() - 类 中的方法org.apache.spark.ml.tuning.CrossValidatorModel
-
- subModels() - 类 中的方法org.apache.spark.ml.tuning.TrainValidationSplitModel
-
- subqueries() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- subqueries() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- subqueries() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- subqueries() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- subsamplingRate() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- subsamplingRate() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- subsamplingRate() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- subsamplingRate() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- subsamplingRate() - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- subsamplingRate() - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- subsamplingRate() - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- subsamplingRate() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- subsamplingRate() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- subsamplingRate() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- subsamplingRate() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- subsamplingRate() - 类 中的方法org.apache.spark.mllib.tree.configuration.Strategy
-
- subsetAccuracy() - 类 中的方法org.apache.spark.mllib.evaluation.MultilabelMetrics
-
Returns subset accuracy
(for equal sets of labels)
- substr(Column, Column) - 类 中的方法org.apache.spark.sql.Column
-
An expression that returns a substring.
- substr(int, int) - 类 中的方法org.apache.spark.sql.Column
-
An expression that returns a substring.
- substring(Column, int, int) - 类 中的静态方法org.apache.spark.sql.functions
-
Substring starts at pos and is of length len when str is String type or
returns the slice of byte array that starts at pos in byte and is of length len
when str is Binary type
- substring_index(Column, String, int) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns the substring from string str before count occurrences of the delimiter delim.
- subtract(JavaDoubleRDD) - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
Return an RDD with the elements from this that are not in other.
- subtract(JavaDoubleRDD, int) - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
Return an RDD with the elements from this that are not in other.
- subtract(JavaDoubleRDD, Partitioner) - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
Return an RDD with the elements from this that are not in other.
- subtract(JavaPairRDD<K, V>) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Return an RDD with the elements from this that are not in other.
- subtract(JavaPairRDD<K, V>, int) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Return an RDD with the elements from this that are not in other.
- subtract(JavaPairRDD<K, V>, Partitioner) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Return an RDD with the elements from this that are not in other.
- subtract(JavaRDD<T>) - 类 中的方法org.apache.spark.api.java.JavaRDD
-
Return an RDD with the elements from this that are not in other.
- subtract(JavaRDD<T>, int) - 类 中的方法org.apache.spark.api.java.JavaRDD
-
Return an RDD with the elements from this that are not in other.
- subtract(JavaRDD<T>, Partitioner) - 类 中的方法org.apache.spark.api.java.JavaRDD
-
Return an RDD with the elements from this that are not in other.
- subtract(RDD<T>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- subtract(RDD<T>, int) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- subtract(RDD<T>, Partitioner, Ordering<T>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- subtract(RDD<T>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- subtract(RDD<T>, int) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- subtract(RDD<T>, Partitioner, Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- subtract(RDD<T>) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- subtract(RDD<T>, int) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- subtract(RDD<T>, Partitioner, Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- subtract(RDD<T>) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- subtract(RDD<T>, int) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- subtract(RDD<T>, Partitioner, Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- subtract(RDD<T>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- subtract(RDD<T>, int) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- subtract(RDD<T>, Partitioner, Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- subtract(BlockMatrix) - 类 中的方法org.apache.spark.mllib.linalg.distributed.BlockMatrix
-
Subtracts the given block matrix other from this block matrix: this - other.
- subtract(RDD<T>) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- subtract(RDD<T>, int) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- subtract(RDD<T>, Partitioner, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- subtract(RDD<T>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- subtract(RDD<T>, int) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- subtract(RDD<T>, Partitioner, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- subtract(RDD<T>) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- subtract(RDD<T>, int) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- subtract(RDD<T>, Partitioner, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- subtract(RDD<T>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- subtract(RDD<T>, int) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- subtract(RDD<T>, Partitioner, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- subtract(RDD<T>) - 类 中的方法org.apache.spark.rdd.RDD
-
Return an RDD with the elements from this that are not in other.
- subtract(RDD<T>, int) - 类 中的方法org.apache.spark.rdd.RDD
-
Return an RDD with the elements from this that are not in other.
- subtract(RDD<T>, Partitioner, Ordering<T>) - 类 中的方法org.apache.spark.rdd.RDD
-
Return an RDD with the elements from this that are not in other.
- subtract(RDD<T>) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- subtract(RDD<T>, int) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- subtract(RDD<T>, Partitioner, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- subtract(long, long) - 类 中的静态方法org.apache.spark.streaming.util.RawTextHelper
-
- subtract$default$3(RDD<T>, Partitioner) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- subtract$default$3(RDD<T>, Partitioner) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- subtract$default$3(RDD<T>, Partitioner) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- subtract$default$3(RDD<T>, Partitioner) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- subtract$default$3(RDD<T>, Partitioner) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- subtract$default$3(RDD<T>, Partitioner) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- subtract$default$3(RDD<T>, Partitioner) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- subtract$default$3(RDD<T>, Partitioner) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- subtract$default$3(RDD<T>, Partitioner) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- subtract$default$3(RDD<T>, Partitioner) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- subtractByKey(JavaPairRDD<K, W>) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Return an RDD with the pairs from this whose keys are not in other.
- subtractByKey(JavaPairRDD<K, W>, int) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Return an RDD with the pairs from this whose keys are not in other.
- subtractByKey(JavaPairRDD<K, W>, Partitioner) - 类 中的方法org.apache.spark.api.java.JavaPairRDD
-
Return an RDD with the pairs from this whose keys are not in other.
- subtractByKey(RDD<Tuple2<K, W>>, ClassTag<W>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Return an RDD with the pairs from this whose keys are not in other.
- subtractByKey(RDD<Tuple2<K, W>>, int, ClassTag<W>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Return an RDD with the pairs from this whose keys are not in other.
- subtractByKey(RDD<Tuple2<K, W>>, Partitioner, ClassTag<W>) - 类 中的方法org.apache.spark.rdd.PairRDDFunctions
-
Return an RDD with the pairs from this whose keys are not in other.
- subtractMetrics(TaskMetrics, TaskMetrics) - 类 中的静态方法org.apache.spark.status.LiveEntityHelpers
-
Subtract m2 values from m1.
- subtreeToString$default$1() - 类 中的静态方法org.apache.spark.ml.tree.InternalNode
-
- succeededTasks() - 类 中的方法org.apache.spark.status.api.v1.ExecutorStageSummary
-
- succeededTasks() - 类 中的方法org.apache.spark.status.LiveExecutorStageSummary
-
- success(T) - 类 中的静态方法org.apache.spark.ml.feature.RFormulaParser
-
- Success - org.apache.spark中的类
-
:: DeveloperApi ::
Task succeeded.
- Success() - 类 的构造器org.apache.spark.Success
-
- successful() - 类 中的方法org.apache.spark.scheduler.TaskInfo
-
- sum() - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
Add up the elements in this RDD.
- Sum() - 类 中的静态方法org.apache.spark.mllib.tree.configuration.EnsembleCombiningStrategy
-
- sum() - 类 中的方法org.apache.spark.rdd.DoubleRDDFunctions
-
Add up the elements in this RDD.
- sum(MapFunction<T, Double>) - 类 中的静态方法org.apache.spark.sql.expressions.javalang.typed
-
Sum aggregate function for floating point (double) type.
- sum(Function1<IN, Object>) - 类 中的静态方法org.apache.spark.sql.expressions.scalalang.typed
-
Sum aggregate function for floating point (double) type.
- sum(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the sum of all values in the expression.
- sum(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the sum of all values in the given column.
- sum(String...) - 类 中的方法org.apache.spark.sql.RelationalGroupedDataset
-
Compute the sum for each numeric columns for each group.
- sum(Seq<String>) - 类 中的方法org.apache.spark.sql.RelationalGroupedDataset
-
Compute the sum for each numeric columns for each group.
- sum(Numeric<B>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- sum() - 类 中的方法org.apache.spark.util.DoubleAccumulator
-
Returns the sum of elements added to the accumulator.
- sum() - 类 中的方法org.apache.spark.util.LongAccumulator
-
Returns the sum of elements added to the accumulator.
- sum() - 类 中的方法org.apache.spark.util.StatCounter
-
- sumApprox(long, Double) - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
Approximate operation to return the sum within a timeout.
- sumApprox(long) - 类 中的方法org.apache.spark.api.java.JavaDoubleRDD
-
Approximate operation to return the sum within a timeout.
- sumApprox(long, double) - 类 中的方法org.apache.spark.rdd.DoubleRDDFunctions
-
Approximate operation to return the sum within a timeout.
- sumDistinct(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the sum of distinct values in the expression.
- sumDistinct(String) - 类 中的静态方法org.apache.spark.sql.functions
-
Aggregate function: returns the sum of distinct values in the expression.
- sumLong(MapFunction<T, Long>) - 类 中的静态方法org.apache.spark.sql.expressions.javalang.typed
-
Sum aggregate function for integral (long, i.e. 64 bit integer) type.
- sumLong(Function1<IN, Object>) - 类 中的静态方法org.apache.spark.sql.expressions.scalalang.typed
-
Sum aggregate function for integral (long, i.e. 64 bit integer) type.
- Summarizer - org.apache.spark.ml.stat中的类
-
Tools for vectorized statistics on MLlib Vectors.
- Summarizer() - 类 的构造器org.apache.spark.ml.stat.Summarizer
-
- summary() - 类 中的方法org.apache.spark.ml.classification.LogisticRegressionModel
-
Gets summary of model on training set.
- summary() - 类 中的方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
Gets summary of model on training set.
- summary() - 类 中的方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
Gets summary of model on training set.
- summary() - 类 中的方法org.apache.spark.ml.clustering.KMeansModel
-
Gets summary of model on training set.
- summary() - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
Gets R-like summary of model on training set.
- summary() - 类 中的方法org.apache.spark.ml.regression.LinearRegressionModel
-
Gets summary (e.g. residuals, mse, r-squared ) of model on training set.
- summary(Column, Column) - 类 中的方法org.apache.spark.ml.stat.SummaryBuilder
-
Returns an aggregate object that contains the summary of the column with the requested metrics.
- summary(Column) - 类 中的方法org.apache.spark.ml.stat.SummaryBuilder
-
- summary(String...) - 类 中的方法org.apache.spark.sql.Dataset
-
Computes specified statistics for numeric and string columns.
- summary(Seq<String>) - 类 中的方法org.apache.spark.sql.Dataset
-
Computes specified statistics for numeric and string columns.
- SummaryBuilder - org.apache.spark.ml.stat中的类
-
A builder object that provides summary statistics about a given column.
- SummaryBuilder() - 类 的构造器org.apache.spark.ml.stat.SummaryBuilder
-
- supportBatch(SparkSession, StructType) - 类 中的静态方法org.apache.spark.sql.hive.orc.OrcFileFormat
-
- supportedFeatureSubsetStrategies() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
Accessor for supported featureSubsetStrategy settings: auto, all, onethird, sqrt, log2
- supportedFeatureSubsetStrategies() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
Accessor for supported featureSubsetStrategy settings: auto, all, onethird, sqrt, log2
- supportedFeatureSubsetStrategies() - 类 中的静态方法org.apache.spark.mllib.tree.RandomForest
-
List of supported feature subset sampling strategies.
- supportedImpurities() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
Accessor for supported impurities: entropy, gini
- supportedImpurities() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
Accessor for supported impurity settings: entropy, gini
- supportedImpurities() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
Accessor for supported impurities: variance
- supportedImpurities() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
Accessor for supported impurity settings: variance
- supportedLossTypes() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
Accessor for supported loss settings: logistic
- supportedLossTypes() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
Accessor for supported loss settings: squared (L2), absolute (L1)
- supportedOptimizers() - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- supportedOptimizers() - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- supportedOptimizers() - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- supportedSelectorTypes() - 类 中的静态方法org.apache.spark.mllib.feature.ChiSqSelector
-
Set of selector types that ChiSqSelector supports.
- SupportsPushDownCatalystFilters - org.apache.spark.sql.sources.v2.reader中的接口
-
- SupportsPushDownFilters - org.apache.spark.sql.sources.v2.reader中的接口
-
- SupportsPushDownRequiredColumns - org.apache.spark.sql.sources.v2.reader中的接口
-
- SupportsReportPartitioning - org.apache.spark.sql.sources.v2.reader中的接口
-
- SupportsReportStatistics - org.apache.spark.sql.sources.v2.reader中的接口
-
- SupportsScanColumnarBatch - org.apache.spark.sql.sources.v2.reader中的接口
-
- SupportsScanUnsafeRow - org.apache.spark.sql.sources.v2.reader中的接口
-
- SupportsWriteInternalRow - org.apache.spark.sql.sources.v2.writer中的接口
-
- surrogateDF() - 类 中的方法org.apache.spark.ml.feature.ImputerModel
-
- SVDPlusPlus - org.apache.spark.graphx.lib中的类
-
Implementation of SVD++ algorithm.
- SVDPlusPlus() - 类 的构造器org.apache.spark.graphx.lib.SVDPlusPlus
-
- SVDPlusPlus.Conf - org.apache.spark.graphx.lib中的类
-
Configuration parameters for SVDPlusPlus.
- SVMDataGenerator - org.apache.spark.mllib.util中的类
-
:: DeveloperApi ::
Generate sample data used for SVM.
- SVMDataGenerator() - 类 的构造器org.apache.spark.mllib.util.SVMDataGenerator
-
- SVMModel - org.apache.spark.mllib.classification中的类
-
Model for Support Vector Machines (SVMs).
- SVMModel(Vector, double) - 类 的构造器org.apache.spark.mllib.classification.SVMModel
-
- SVMWithSGD - org.apache.spark.mllib.classification中的类
-
Train a Support Vector Machine (SVM) using Stochastic Gradient Descent.
- SVMWithSGD() - 类 的构造器org.apache.spark.mllib.classification.SVMWithSGD
-
Construct a SVM object with default parameters: {stepSize: 1.0, numIterations: 100,
regParm: 0.01, miniBatchFraction: 1.0}.
- symbolToColumn(Symbol) - 类 中的方法org.apache.spark.sql.SQLImplicits
-
An implicit conversion that turns a Scala
Symbol into a
Column.
- symlink(File, File) - 类 中的静态方法org.apache.spark.util.Utils
-
Creates a symlink.
- symmetricEigs(Function1<DenseVector<Object>, DenseVector<Object>>, int, int, double, int) - 类 中的静态方法org.apache.spark.mllib.linalg.EigenValueDecomposition
-
Compute the leading k eigenvalues and eigenvectors on a symmetric square matrix using ARPACK.
- syr(double, Vector, DenseMatrix) - 类 中的静态方法org.apache.spark.ml.linalg.BLAS
-
A := alpha * x * x^T^ + A
- syr(double, Vector, DenseMatrix) - 类 中的静态方法org.apache.spark.mllib.linalg.BLAS
-
A := alpha * x * x^T^ + A
- SYSTEM_DEFAULT() - 类 中的静态方法org.apache.spark.sql.types.DecimalType
-
- systemProperties() - 类 中的方法org.apache.spark.status.api.v1.ApplicationEnvironmentInfo
-
- t() - 类 中的方法org.apache.spark.SerializableWritable
-
- Table - org.apache.spark.sql.catalog中的类
-
A table in Spark, as returned by the
listTables method in
Catalog.
- Table(String, String, String, String, boolean) - 类 的构造器org.apache.spark.sql.catalog.Table
-
- table(String) - 类 中的方法org.apache.spark.sql.DataFrameReader
-
Returns the specified table as a DataFrame.
- table() - 类 中的方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- table(String) - 类 中的方法org.apache.spark.sql.SparkSession
-
Returns the specified table/view as a DataFrame.
- table(String) - 类 中的方法org.apache.spark.sql.SQLContext
-
- TABLE_CLASS_NOT_STRIPED() - 类 中的静态方法org.apache.spark.ui.UIUtils
-
- TABLE_CLASS_STRIPED() - 类 中的静态方法org.apache.spark.ui.UIUtils
-
- TABLE_CLASS_STRIPED_SORTABLE() - 类 中的静态方法org.apache.spark.ui.UIUtils
-
- tableDesc() - 类 中的方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- tableExists(String) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
Check if the table or view with the specified name exists.
- tableExists(String, String) - 类 中的方法org.apache.spark.sql.catalog.Catalog
-
Check if the table or view with the specified name exists in the specified database.
- tableNames() - 类 中的方法org.apache.spark.sql.SQLContext
-
- tableNames(String) - 类 中的方法org.apache.spark.sql.SQLContext
-
- tables() - 类 中的方法org.apache.spark.sql.SQLContext
-
- tables(String) - 类 中的方法org.apache.spark.sql.SQLContext
-
- TableScan - org.apache.spark.sql.sources中的接口
-
A BaseRelation that can produce all of its tuples as an RDD of Row objects.
- tableType() - 类 中的方法org.apache.spark.sql.catalog.Table
-
- tail() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- tails() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- take(int) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- take(int) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- take(int) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- take(int) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Take the first num elements of the RDD.
- take(int) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- take(int) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- take(int) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- take(int) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- take(int) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- take(int) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- take(int) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- take(int) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- take(int) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- take(int) - 类 中的方法org.apache.spark.rdd.RDD
-
Take the first num elements of the RDD.
- take(int) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- take(int) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns the first n rows in the Dataset.
- take(int) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- takeAsList(int) - 类 中的方法org.apache.spark.sql.Dataset
-
Returns the first n rows in the Dataset as a list.
- takeAsync(int) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- takeAsync(int) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- takeAsync(int) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- takeAsync(int) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
The asynchronous version of the take action, which returns a
future for retrieving the first num elements of this RDD.
- takeAsync(int) - 类 中的方法org.apache.spark.rdd.AsyncRDDActions
-
Returns a future for retrieving the first num elements of the RDD.
- takeOrdered(int, Comparator<T>) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- takeOrdered(int) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- takeOrdered(int, Comparator<T>) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- takeOrdered(int) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- takeOrdered(int, Comparator<T>) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- takeOrdered(int) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- takeOrdered(int, Comparator<T>) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Returns the first k (smallest) elements from this RDD as defined by
the specified Comparator[T] and maintains the order.
- takeOrdered(int) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Returns the first k (smallest) elements from this RDD using the
natural ordering for T while maintain the order.
- takeOrdered(int, Ordering<T>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- takeOrdered(int, Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- takeOrdered(int, Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- takeOrdered(int, Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- takeOrdered(int, Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- takeOrdered(int, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- takeOrdered(int, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- takeOrdered(int, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- takeOrdered(int, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- takeOrdered(int, Ordering<T>) - 类 中的方法org.apache.spark.rdd.RDD
-
Returns the first k (smallest) elements from this RDD as defined by the specified
implicit Ordering[T] and maintains the ordering.
- takeOrdered(int, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- takeRight(int) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- takeSample(boolean, int) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- takeSample(boolean, int, long) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- takeSample(boolean, int) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- takeSample(boolean, int, long) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- takeSample(boolean, int) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- takeSample(boolean, int, long) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- takeSample(boolean, int) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
- takeSample(boolean, int, long) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
- takeSample(boolean, int, long) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- takeSample(boolean, int, long) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- takeSample(boolean, int, long) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- takeSample(boolean, int, long) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- takeSample(boolean, int, long) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- takeSample(boolean, int, long) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- takeSample(boolean, int, long) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- takeSample(boolean, int, long) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- takeSample(boolean, int, long) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- takeSample(boolean, int, long) - 类 中的方法org.apache.spark.rdd.RDD
-
Return a fixed-size sampled subset of this RDD in an array
- takeSample(boolean, int, long) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- takeSample$default$3() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- takeSample$default$3() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- takeSample$default$3() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- takeSample$default$3() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- takeSample$default$3() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- takeSample$default$3() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- takeSample$default$3() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- takeSample$default$3() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- takeSample$default$3() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- takeSample$default$3() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- takeWhile(Function1<A, Object>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- tallSkinnyQR(boolean) - 类 中的方法org.apache.spark.mllib.linalg.distributed.RowMatrix
-
- tan(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
- tan(String) - 类 中的静态方法org.apache.spark.sql.functions
-
- tanh(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
- tanh(String) - 类 中的静态方法org.apache.spark.sql.functions
-
- targetStorageLevel() - 类 中的方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- targetStorageLevel() - 类 中的方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- task() - 类 中的方法org.apache.spark.CleanupTaskWeakReference
-
- TASK_DESERIALIZATION_TIME() - 类 中的静态方法org.apache.spark.ui.jobs.TaskDetailsClassNames
-
- TASK_DESERIALIZATION_TIME() - 类 中的静态方法org.apache.spark.ui.ToolTips
-
- TASK_INDEX() - 类 中的静态方法org.apache.spark.status.TaskIndexNames
-
- TASK_TIME() - 类 中的静态方法org.apache.spark.ui.ToolTips
-
- taskAttemptId() - 类 中的方法org.apache.spark.TaskContext
-
An ID that is unique to this task attempt (within the same SparkContext, no two task attempts
will share the same attempt ID).
- TaskCommitDenied - org.apache.spark中的类
-
:: DeveloperApi ::
Task requested the driver to commit, but was denied.
- TaskCommitDenied(int, int, int) - 类 的构造器org.apache.spark.TaskCommitDenied
-
- TaskCommitMessage(Object) - 类 的构造器org.apache.spark.internal.io.FileCommitProtocol.TaskCommitMessage
-
- TaskCompletionListener - org.apache.spark.util中的接口
-
:: DeveloperApi ::
Listener providing a callback function to invoke when a task's execution completes.
- TaskContext - org.apache.spark中的类
-
Contextual information about a task which can be read or mutated during
execution.
- TaskContext() - 类 的构造器org.apache.spark.TaskContext
-
- taskCount(StageData) - 类 中的静态方法org.apache.spark.ui.jobs.ApiHelper
-
- TaskData - org.apache.spark.status.api.v1中的类
-
- TaskDetailsClassNames - org.apache.spark.ui.jobs中的类
-
Names of the CSS classes corresponding to each type of task detail.
- TaskDetailsClassNames() - 类 的构造器org.apache.spark.ui.jobs.TaskDetailsClassNames
-
- taskEndFromJson(JsonAST.JValue) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- TaskEndReason - org.apache.spark中的接口
-
:: DeveloperApi ::
Various possible reasons why a task ended.
- taskEndReasonFromJson(JsonAST.JValue) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- taskEndReasonToJson(TaskEndReason) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- taskEndToJson(SparkListenerTaskEnd) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- TaskFailedReason - org.apache.spark中的接口
-
:: DeveloperApi ::
Various possible reasons why a task failed.
- TaskFailureListener - org.apache.spark.util中的接口
-
:: DeveloperApi ::
Listener providing a callback function to invoke when a task's execution encounters an error.
- taskFailures() - 类 中的方法org.apache.spark.scheduler.SparkListenerExecutorBlacklisted
-
- taskGettingResultFromJson(JsonAST.JValue) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- taskGettingResultToJson(SparkListenerTaskGettingResult) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- taskId() - 类 中的方法org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.KillTask
-
- taskId() - 类 中的方法org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.StatusUpdate
-
- taskId() - 类 中的方法org.apache.spark.scheduler.local.KillTask
-
- taskId() - 类 中的方法org.apache.spark.scheduler.local.StatusUpdate
-
- taskId() - 类 中的方法org.apache.spark.scheduler.TaskInfo
-
- taskId() - 类 中的方法org.apache.spark.status.api.v1.TaskData
-
- taskId() - 类 中的方法org.apache.spark.storage.TaskResultBlockId
-
- TaskIndexNames - org.apache.spark.status中的类
-
Tasks have a lot of indices that are used in a few different places.
- TaskIndexNames() - 类 的构造器org.apache.spark.status.TaskIndexNames
-
- taskInfo() - 类 中的方法org.apache.spark.scheduler.SparkListenerTaskEnd
-
- taskInfo() - 类 中的方法org.apache.spark.scheduler.SparkListenerTaskGettingResult
-
- taskInfo() - 类 中的方法org.apache.spark.scheduler.SparkListenerTaskStart
-
- TaskInfo - org.apache.spark.scheduler中的类
-
:: DeveloperApi ::
Information about a running task attempt inside a TaskSet.
- TaskInfo(long, int, int, long, String, String, Enumeration.Value, boolean) - 类 的构造器org.apache.spark.scheduler.TaskInfo
-
- taskInfoFromJson(JsonAST.JValue) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- taskInfoToJson(TaskInfo) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- TaskKilled - org.apache.spark中的类
-
:: DeveloperApi ::
Task was killed intentionally and needs to be rescheduled.
- TaskKilled(String) - 类 的构造器org.apache.spark.TaskKilled
-
- TaskKilledException - org.apache.spark中的异常错误
-
:: DeveloperApi ::
Exception thrown when a task is explicitly killed (i.e., task failure is expected).
- TaskKilledException(String) - 异常错误 的构造器org.apache.spark.TaskKilledException
-
- TaskKilledException() - 异常错误 的构造器org.apache.spark.TaskKilledException
-
- taskLocality() - 类 中的方法org.apache.spark.scheduler.TaskInfo
-
- TaskLocality - org.apache.spark.scheduler中的类
-
- TaskLocality() - 类 的构造器org.apache.spark.scheduler.TaskLocality
-
- taskLocality() - 类 中的方法org.apache.spark.status.api.v1.TaskData
-
- TaskMetricDistributions - org.apache.spark.status.api.v1中的类
-
- taskMetrics() - 类 中的方法org.apache.spark.scheduler.SparkListenerTaskEnd
-
- taskMetrics() - 类 中的方法org.apache.spark.scheduler.StageInfo
-
- taskMetrics() - 类 中的方法org.apache.spark.status.api.v1.TaskData
-
- TaskMetrics - org.apache.spark.status.api.v1中的类
-
- taskMetrics() - 类 中的方法org.apache.spark.TaskContext
-
- taskMetricsFromJson(JsonAST.JValue) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- taskMetricsToJson(TaskMetrics) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- TASKRESULT() - 类 中的静态方法org.apache.spark.storage.BlockId
-
- TaskResultBlockId - org.apache.spark.storage中的类
-
- TaskResultBlockId(long) - 类 的构造器org.apache.spark.storage.TaskResultBlockId
-
- TaskResultLost - org.apache.spark中的类
-
:: DeveloperApi ::
The task finished successfully, but the result was lost from the executor's block manager before
it was fetched.
- TaskResultLost() - 类 的构造器org.apache.spark.TaskResultLost
-
- tasks() - 类 中的方法org.apache.spark.status.api.v1.StageData
-
- TaskSchedulerIsSet - org.apache.spark中的类
-
An event that SparkContext uses to notify HeartbeatReceiver that SparkContext.taskScheduler is
created.
- TaskSchedulerIsSet() - 类 的构造器org.apache.spark.TaskSchedulerIsSet
-
- TaskSorting - org.apache.spark.status.api.v1中的枚举
-
- taskStartFromJson(JsonAST.JValue) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- taskStartToJson(SparkListenerTaskStart) - 类 中的静态方法org.apache.spark.util.JsonProtocol
-
- TaskState - org.apache.spark中的类
-
- TaskState() - 类 的构造器org.apache.spark.TaskState
-
- taskTime() - 类 中的方法org.apache.spark.status.api.v1.ExecutorStageSummary
-
- taskTime() - 类 中的方法org.apache.spark.status.LiveExecutorStageSummary
-
- taskType() - 类 中的方法org.apache.spark.scheduler.SparkListenerTaskEnd
-
- TEMP_DIR_SHUTDOWN_PRIORITY() - 类 中的静态方法org.apache.spark.util.ShutdownHookManager
-
The shutdown priority of temp directory must be lower than the SparkContext shutdown
priority.
- TEMP_LOCAL() - 类 中的静态方法org.apache.spark.storage.BlockId
-
- TEMP_SHUFFLE() - 类 中的静态方法org.apache.spark.storage.BlockId
-
- tempFileWith(File) - 类 中的静态方法org.apache.spark.util.Utils
-
Returns a path of temporary file which is in the same directory with path.
- TeradataDialect - org.apache.spark.sql.jdbc中的类
-
- TeradataDialect() - 类 的构造器org.apache.spark.sql.jdbc.TeradataDialect
-
- terminateProcess(Process, long) - 类 中的静态方法org.apache.spark.util.Utils
-
Terminates a process waiting for at most the specified duration.
- test(Dataset<Row>, String, String) - 类 中的静态方法org.apache.spark.ml.stat.ChiSquareTest
-
Conduct Pearson's independence test for every feature against the label.
- TEST() - 类 中的静态方法org.apache.spark.storage.BlockId
-
- TEST_ACCUM() - 类 中的静态方法org.apache.spark.InternalAccumulator
-
- testCommandAvailable(String) - 类 中的静态方法org.apache.spark.TestUtils
-
Test if a command is available.
- testOneSample(RDD<Object>, String, double...) - 类 中的静态方法org.apache.spark.mllib.stat.test.KolmogorovSmirnovTest
-
A convenience function that allows running the KS test for 1 set of sample data against
a named distribution
- testOneSample(RDD<Object>, Function1<Object, Object>) - 类 中的静态方法org.apache.spark.mllib.stat.test.KolmogorovSmirnovTest
-
- testOneSample(RDD<Object>, RealDistribution) - 类 中的静态方法org.apache.spark.mllib.stat.test.KolmogorovSmirnovTest
-
- testOneSample(RDD<Object>, String, Seq<Object>) - 类 中的静态方法org.apache.spark.mllib.stat.test.KolmogorovSmirnovTest
-
- TestResult<DF> - org.apache.spark.mllib.stat.test中的接口
-
Trait for hypothesis test results.
- TestUtils - org.apache.spark中的类
-
Utilities for tests.
- TestUtils() - 类 的构造器org.apache.spark.TestUtils
-
- text(String...) - 类 中的方法org.apache.spark.sql.DataFrameReader
-
Loads text files and returns a DataFrame whose schema starts with a string column named
"value", and followed by partitioned columns if there are any.
- text(String) - 类 中的方法org.apache.spark.sql.DataFrameReader
-
Loads text files and returns a DataFrame whose schema starts with a string column named
"value", and followed by partitioned columns if there are any.
- text(Seq<String>) - 类 中的方法org.apache.spark.sql.DataFrameReader
-
Loads text files and returns a DataFrame whose schema starts with a string column named
"value", and followed by partitioned columns if there are any.
- text(String) - 类 中的方法org.apache.spark.sql.DataFrameWriter
-
Saves the content of the DataFrame in a text file at the specified path.
- text(String) - 类 中的方法org.apache.spark.sql.streaming.DataStreamReader
-
Loads text files and returns a DataFrame whose schema starts with a string column named
"value", and followed by partitioned columns if there are any.
- textFile(String) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
Read a text file from HDFS, a local file system (available on all nodes), or any
Hadoop-supported file system URI, and return it as an RDD of Strings.
- textFile(String, int) - 类 中的方法org.apache.spark.api.java.JavaSparkContext
-
Read a text file from HDFS, a local file system (available on all nodes), or any
Hadoop-supported file system URI, and return it as an RDD of Strings.
- textFile(String, int) - 类 中的方法org.apache.spark.SparkContext
-
Read a text file from HDFS, a local file system (available on all nodes), or any
Hadoop-supported file system URI, and return it as an RDD of Strings.
- textFile(String...) - 类 中的方法org.apache.spark.sql.DataFrameReader
-
Loads text files and returns a
Dataset of String.
- textFile(String) - 类 中的方法org.apache.spark.sql.DataFrameReader
-
Loads text files and returns a
Dataset of String.
- textFile(Seq<String>) - 类 中的方法org.apache.spark.sql.DataFrameReader
-
Loads text files and returns a
Dataset of String.
- textFile(String) - 类 中的方法org.apache.spark.sql.streaming.DataStreamReader
-
Loads text file(s) and returns a Dataset of String.
- textFileStream(String) - 类 中的方法org.apache.spark.streaming.api.java.JavaStreamingContext
-
Create an input stream that monitors a Hadoop-compatible filesystem
for new files and reads them as text files (using key as LongWritable, value
as Text and input format as TextInputFormat).
- textFileStream(String) - 类 中的方法org.apache.spark.streaming.StreamingContext
-
Create an input stream that monitors a Hadoop-compatible filesystem
for new files and reads them as text files (using key as LongWritable, value
as Text and input format as TextInputFormat).
- textResponderToServlet(Function1<HttpServletRequest, String>) - 类 中的静态方法org.apache.spark.ui.JettyUtils
-
- theta() - 类 中的方法org.apache.spark.ml.classification.NaiveBayesModel
-
- theta() - 类 中的方法org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV1_0$.Data
-
- theta() - 类 中的方法org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV2_0$.Data
-
- theta() - 类 中的方法org.apache.spark.mllib.classification.NaiveBayesModel
-
- thisClassName() - 类 中的方法org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV1_0$
-
Hard-code class name string in case it changes in the future
- thisClassName() - 类 中的方法org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV2_0$
-
Hard-code class name string in case it changes in the future
- thisClassName() - 类 中的方法org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$
-
- thisFormatVersion() - 类 中的方法org.apache.spark.mllib.classification.impl.GLMClassificationModel.SaveLoadV1_0$
-
- thisFormatVersion() - 类 中的方法org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV1_0$
-
- thisFormatVersion() - 类 中的方法org.apache.spark.mllib.classification.NaiveBayesModel.SaveLoadV2_0$
-
- thisFormatVersion() - 类 中的方法org.apache.spark.mllib.regression.impl.GLMRegressionModel.SaveLoadV1_0$
-
- thisFormatVersion() - 类 中的方法org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$
-
- ThreadUtils - org.apache.spark.util中的类
-
- ThreadUtils() - 类 的构造器org.apache.spark.util.ThreadUtils
-
- threshold() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- threshold() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- threshold() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- threshold() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- threshold() - 类 中的方法org.apache.spark.ml.feature.Binarizer
-
Param for threshold used to binarize continuous features.
- threshold() - 接口 中的方法org.apache.spark.ml.param.shared.HasThreshold
-
Param for threshold in binary classification prediction, in range [0, 1].
- threshold() - 类 中的方法org.apache.spark.ml.tree.ContinuousSplit
-
- threshold() - 类 中的方法org.apache.spark.mllib.classification.impl.GLMClassificationModel.SaveLoadV1_0$.Data
-
- threshold() - 类 中的方法org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.SplitData
-
- threshold() - 类 中的方法org.apache.spark.mllib.tree.model.Split
-
- thresholds() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- thresholds() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- thresholds() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- thresholds() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- thresholds() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- thresholds() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- thresholds() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- thresholds() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- thresholds() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- thresholds() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- thresholds() - 类 中的静态方法org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- thresholds() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- thresholds() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- thresholds() - 接口 中的方法org.apache.spark.ml.param.shared.HasThresholds
-
Param for Thresholds in multi-class classification to adjust the probability of predicting each class.
- thresholds() - 类 中的方法org.apache.spark.mllib.evaluation.BinaryClassificationMetrics
-
Returns thresholds in descending order.
- throwBalls(int, RDD<?>, double, DefaultPartitionCoalescer.PartitionLocations) - 类 中的方法org.apache.spark.rdd.DefaultPartitionCoalescer
-
- time() - 类 中的方法org.apache.spark.scheduler.SparkListenerApplicationEnd
-
- time() - 类 中的方法org.apache.spark.scheduler.SparkListenerApplicationStart
-
- time() - 类 中的方法org.apache.spark.scheduler.SparkListenerBlockManagerAdded
-
- time() - 类 中的方法org.apache.spark.scheduler.SparkListenerBlockManagerRemoved
-
- time() - 类 中的方法org.apache.spark.scheduler.SparkListenerExecutorAdded
-
- time() - 类 中的方法org.apache.spark.scheduler.SparkListenerExecutorBlacklisted
-
- time() - 类 中的方法org.apache.spark.scheduler.SparkListenerExecutorRemoved
-
- time() - 类 中的方法org.apache.spark.scheduler.SparkListenerExecutorUnblacklisted
-
- time() - 类 中的方法org.apache.spark.scheduler.SparkListenerJobEnd
-
- time() - 类 中的方法org.apache.spark.scheduler.SparkListenerJobStart
-
- time() - 类 中的方法org.apache.spark.scheduler.SparkListenerNodeBlacklisted
-
- time() - 类 中的方法org.apache.spark.scheduler.SparkListenerNodeUnblacklisted
-
- time(Function0<T>) - 类 中的方法org.apache.spark.sql.SparkSession
-
Executes some code block and prints to stdout the time taken to execute the block.
- time() - 异常错误 中的方法org.apache.spark.sql.streaming.StreamingQueryException
-
Time when the exception occurred
- time() - 类 中的方法org.apache.spark.streaming.scheduler.StreamingListenerStreamingStarted
-
- Time - org.apache.spark.streaming中的类
-
This is a simple class that represents an absolute instant of time.
- Time(long) - 类 的构造器org.apache.spark.streaming.Time
-
- timeFromString(String, TimeUnit) - 类 中的静态方法org.apache.spark.internal.config.ConfigHelpers
-
- timeIt(int, Function0<BoxedUnit>, Option<Function0<BoxedUnit>>) - 类 中的静态方法org.apache.spark.util.Utils
-
Timing method based on iterations that permit JVM JIT optimization.
- timeout(Duration) - 类 中的方法org.apache.spark.streaming.StateSpec
-
Set the duration after which the state of an idle key will be removed.
- TIMER() - 类 中的静态方法org.apache.spark.metrics.sink.StatsdMetricType
-
- times(int) - 类 中的方法org.apache.spark.streaming.Duration
-
- times(int, Function0<BoxedUnit>) - 类 中的静态方法org.apache.spark.util.Utils
-
Method executed for repeating a task for side effects.
- timestamp() - 类 中的方法org.apache.spark.sql.ColumnName
-
Creates a new StructField of type timestamp.
- TIMESTAMP() - 类 中的静态方法org.apache.spark.sql.Encoders
-
An encoder for nullable timestamp type.
- timestamp() - 类 中的方法org.apache.spark.sql.streaming.StreamingQueryProgress
-
- TimestampType - 类 中的静态变量org.apache.spark.sql.types.DataTypes
-
Gets the TimestampType object.
- TimestampType - org.apache.spark.sql.types中的类
-
The data type representing java.sql.Timestamp values.
- timeStringAsMs(String) - 类 中的静态方法org.apache.spark.util.Utils
-
Convert a time parameter such as (50s, 100ms, or 250us) to microseconds for internal use.
- timeStringAsSeconds(String) - 类 中的静态方法org.apache.spark.util.Utils
-
Convert a time parameter such as (50s, 100ms, or 250us) to seconds for internal use.
- timeToString(long, TimeUnit) - 类 中的静态方法org.apache.spark.internal.config.ConfigHelpers
-
- TimeTrackingOutputStream - org.apache.spark.storage中的类
-
Intercepts write calls and tracks total time spent writing in order to update shuffle write
metrics.
- TimeTrackingOutputStream(ShuffleWriteMetrics, OutputStream) - 类 的构造器org.apache.spark.storage.TimeTrackingOutputStream
-
- timeUnit() - 类 中的方法org.apache.spark.mllib.clustering.StreamingKMeans
-
- TIMING_DATA() - 类 中的静态方法org.apache.spark.api.r.SpecialLengths
-
- to(CanBuildFrom<Nothing$, A, Col>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- to(Time, Duration) - 类 中的方法org.apache.spark.streaming.Time
-
- to_date(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Converts the column into DateType by casting rules to DateType.
- to_date(Column, String) - 类 中的静态方法org.apache.spark.sql.functions
-
Converts the column into a DateType with a specified format
(see [http://docs.oracle.com/javase/tutorial/i18n/format/simpleDateFormat.html])
return null if fail.
- to_json(Column, Map<String, String>) - 类 中的静态方法org.apache.spark.sql.functions
-
(Scala-specific) Converts a column containing a StructType, ArrayType of StructTypes,
a MapType or ArrayType of MapTypes into a JSON string with the specified schema.
- to_json(Column, Map<String, String>) - 类 中的静态方法org.apache.spark.sql.functions
-
(Java-specific) Converts a column containing a StructType, ArrayType of StructTypes,
a MapType or ArrayType of MapTypes into a JSON string with the specified schema.
- to_json(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Converts a column containing a StructType, ArrayType of StructTypes,
a MapType or ArrayType of MapTypes into a JSON string with the specified schema.
- to_timestamp(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Convert time string to a Unix timestamp (in seconds) by casting rules to TimestampType.
- to_timestamp(Column, String) - 类 中的静态方法org.apache.spark.sql.functions
-
Convert time string to a Unix timestamp (in seconds) with a specified format
(see [http://docs.oracle.com/javase/tutorial/i18n/format/simpleDateFormat.html])
to Unix timestamp (in seconds), return null if fail.
- to_utc_timestamp(Column, String) - 类 中的静态方法org.apache.spark.sql.functions
-
Given a timestamp like '2017-07-14 02:40:00.0', interprets it as a time in the given time
zone, and renders that time as a timestamp in UTC.
- toApi() - 类 中的方法org.apache.spark.status.LiveRDDDistribution
-
- toApi() - 类 中的方法org.apache.spark.status.LiveStage
-
- toArray() - 类 中的方法org.apache.spark.input.PortableDataStream
-
Read the file as a byte array
- toArray() - 类 中的静态方法org.apache.spark.ml.linalg.DenseMatrix
-
- toArray() - 类 中的方法org.apache.spark.ml.linalg.DenseVector
-
- toArray() - 接口 中的方法org.apache.spark.ml.linalg.Matrix
-
Converts to a dense array in column major.
- toArray() - 类 中的静态方法org.apache.spark.ml.linalg.SparseMatrix
-
- toArray() - 类 中的方法org.apache.spark.ml.linalg.SparseVector
-
- toArray() - 接口 中的方法org.apache.spark.ml.linalg.Vector
-
Converts the instance to a double array.
- toArray() - 类 中的静态方法org.apache.spark.mllib.linalg.DenseMatrix
-
- toArray() - 类 中的方法org.apache.spark.mllib.linalg.DenseVector
-
- toArray() - 接口 中的方法org.apache.spark.mllib.linalg.Matrix
-
Converts to a dense array in column major.
- toArray() - 类 中的静态方法org.apache.spark.mllib.linalg.SparseMatrix
-
- toArray() - 类 中的方法org.apache.spark.mllib.linalg.SparseVector
-
- toArray() - 接口 中的方法org.apache.spark.mllib.linalg.Vector
-
Converts the instance to a double array.
- toArray(ClassTag<B>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- toBigDecimal() - 类 中的方法org.apache.spark.sql.types.Decimal
-
- toBlockMatrix() - 类 中的方法org.apache.spark.mllib.linalg.distributed.CoordinateMatrix
-
Converts to BlockMatrix.
- toBlockMatrix(int, int) - 类 中的方法org.apache.spark.mllib.linalg.distributed.CoordinateMatrix
-
Converts to BlockMatrix.
- toBlockMatrix() - 类 中的方法org.apache.spark.mllib.linalg.distributed.IndexedRowMatrix
-
Converts to BlockMatrix.
- toBlockMatrix(int, int) - 类 中的方法org.apache.spark.mllib.linalg.distributed.IndexedRowMatrix
-
Converts to BlockMatrix.
- toBoolean(String, String) - 类 中的静态方法org.apache.spark.internal.config.ConfigHelpers
-
- toBooleanArray() - 类 中的方法org.apache.spark.sql.vectorized.ColumnarArray
-
- toBreeze() - 接口 中的方法org.apache.spark.mllib.linalg.distributed.DistributedMatrix
-
Collects data and assembles a local dense breeze matrix (for test only).
- toBuffer() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- toByte() - 类 中的方法org.apache.spark.sql.types.Decimal
-
- toByteArray() - 类 中的方法org.apache.spark.sql.vectorized.ColumnarArray
-
- toByteArray() - 类 中的方法org.apache.spark.util.sketch.CountMinSketch
-
- toByteBuffer() - 类 中的方法org.apache.spark.storage.DiskBlockData
-
- toByteBuffer() - 类 中的方法org.apache.spark.storage.EncryptedBlockData
-
- toCatalystDecimal(HiveDecimalObjectInspector, Object) - 类 中的静态方法org.apache.spark.sql.hive.HiveShim
-
- toChunkedByteBuffer(Function1<Object, ByteBuffer>) - 类 中的方法org.apache.spark.storage.DiskBlockData
-
- toChunkedByteBuffer(Function1<Object, ByteBuffer>) - 类 中的方法org.apache.spark.storage.EncryptedBlockData
-
- toColumn() - 类 中的方法org.apache.spark.sql.expressions.Aggregator
-
Returns this Aggregator as a TypedColumn that can be used in Dataset.
- toCoordinateMatrix() - 类 中的方法org.apache.spark.mllib.linalg.distributed.BlockMatrix
-
Converts to CoordinateMatrix.
- toCoordinateMatrix() - 类 中的方法org.apache.spark.mllib.linalg.distributed.IndexedRowMatrix
-
- toCryptoConf(SparkConf) - 类 中的静态方法org.apache.spark.security.CryptoStreamUtils
-
- toDebugString() - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- toDebugString() - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- toDebugString() - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- toDebugString() - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
A description of this RDD and its recursive dependencies for debugging.
- toDebugString() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- toDebugString() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- toDebugString() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- toDebugString() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- toDebugString() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- toDebugString() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- toDebugString() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- toDebugString() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- toDebugString() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- toDebugString() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- toDebugString() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- toDebugString() - 类 中的方法org.apache.spark.mllib.tree.model.DecisionTreeModel
-
Print the full model to a string.
- toDebugString() - 类 中的静态方法org.apache.spark.mllib.tree.model.GradientBoostedTreesModel
-
- toDebugString() - 类 中的静态方法org.apache.spark.mllib.tree.model.RandomForestModel
-
- toDebugString() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- toDebugString() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- toDebugString() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- toDebugString() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- toDebugString() - 类 中的方法org.apache.spark.rdd.RDD
-
A description of this RDD and its recursive dependencies for debugging.
- toDebugString() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- toDebugString() - 类 中的方法org.apache.spark.SparkConf
-
Return a string listing all keys and values, one per line.
- toDebugString() - 类 中的方法org.apache.spark.sql.types.Decimal
-
- toDegrees(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
- toDegrees(String) - 类 中的静态方法org.apache.spark.sql.functions
-
- toDense() - 类 中的静态方法org.apache.spark.ml.linalg.DenseMatrix
-
- toDense() - 类 中的静态方法org.apache.spark.ml.linalg.DenseVector
-
- toDense() - 接口 中的方法org.apache.spark.ml.linalg.Matrix
-
Converts this matrix to a dense matrix while maintaining the layout of the current matrix.
- toDense() - 类 中的静态方法org.apache.spark.ml.linalg.SparseMatrix
-
- toDense() - 类 中的静态方法org.apache.spark.ml.linalg.SparseVector
-
- toDense() - 接口 中的方法org.apache.spark.ml.linalg.Vector
-
Converts this vector to a dense vector.
- toDense() - 类 中的静态方法org.apache.spark.mllib.linalg.DenseVector
-
- toDense() - 类 中的方法org.apache.spark.mllib.linalg.SparseMatrix
-
Generate a DenseMatrix from the given SparseMatrix.
- toDense() - 类 中的静态方法org.apache.spark.mllib.linalg.SparseVector
-
- toDense() - 接口 中的方法org.apache.spark.mllib.linalg.Vector
-
Converts this vector to a dense vector.
- toDenseColMajor() - 类 中的静态方法org.apache.spark.ml.linalg.DenseMatrix
-
- toDenseColMajor() - 接口 中的方法org.apache.spark.ml.linalg.Matrix
-
Converts this matrix to a dense matrix in column major order.
- toDenseColMajor() - 类 中的静态方法org.apache.spark.ml.linalg.SparseMatrix
-
- toDenseMatrix(boolean) - 接口 中的方法org.apache.spark.ml.linalg.Matrix
-
Converts this matrix to a dense matrix.
- toDenseRowMajor() - 类 中的静态方法org.apache.spark.ml.linalg.DenseMatrix
-
- toDenseRowMajor() - 接口 中的方法org.apache.spark.ml.linalg.Matrix
-
Converts this matrix to a dense matrix in row major order.
- toDenseRowMajor() - 类 中的静态方法org.apache.spark.ml.linalg.SparseMatrix
-
- toDF(String...) - 类 中的方法org.apache.spark.sql.Dataset
-
Converts this strongly typed collection of data to generic DataFrame with columns renamed.
- toDF() - 类 中的方法org.apache.spark.sql.Dataset
-
Converts this strongly typed collection of data to generic Dataframe.
- toDF(Seq<String>) - 类 中的方法org.apache.spark.sql.Dataset
-
Converts this strongly typed collection of data to generic DataFrame with columns renamed.
- toDF() - 类 中的方法org.apache.spark.sql.DatasetHolder
-
- toDF(Seq<String>) - 类 中的方法org.apache.spark.sql.DatasetHolder
-
- toDouble() - 类 中的方法org.apache.spark.sql.types.Decimal
-
- toDoubleArray() - 类 中的方法org.apache.spark.sql.vectorized.ColumnarArray
-
- toDS() - 类 中的方法org.apache.spark.sql.DatasetHolder
-
- toEdgeTriplet() - 类 中的方法org.apache.spark.graphx.EdgeContext
-
Converts the edge and vertex properties into an
EdgeTriplet for convenience.
- toErrorString() - 类 中的方法org.apache.spark.ExceptionFailure
-
- toErrorString() - 类 中的方法org.apache.spark.ExecutorLostFailure
-
- toErrorString() - 类 中的方法org.apache.spark.FetchFailed
-
- toErrorString() - 类 中的静态方法org.apache.spark.Resubmitted
-
- toErrorString() - 类 中的方法org.apache.spark.TaskCommitDenied
-
- toErrorString() - 接口 中的方法org.apache.spark.TaskFailedReason
-
Error message displayed in the web UI.
- toErrorString() - 类 中的方法org.apache.spark.TaskKilled
-
- toErrorString() - 类 中的静态方法org.apache.spark.TaskResultLost
-
- toErrorString() - 类 中的静态方法org.apache.spark.UnknownReason
-
- toFloat() - 类 中的方法org.apache.spark.sql.types.Decimal
-
- toFloatArray() - 类 中的方法org.apache.spark.sql.vectorized.ColumnarArray
-
- toFormattedString() - 类 中的方法org.apache.spark.streaming.Duration
-
- toIndexedRowMatrix() - 类 中的方法org.apache.spark.mllib.linalg.distributed.BlockMatrix
-
Converts to IndexedRowMatrix.
- toIndexedRowMatrix() - 类 中的方法org.apache.spark.mllib.linalg.distributed.CoordinateMatrix
-
Converts to IndexedRowMatrix.
- toIndexedSeq() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- toInputStream() - 类 中的方法org.apache.spark.storage.DiskBlockData
-
- toInputStream() - 类 中的方法org.apache.spark.storage.EncryptedBlockData
-
- toInspector(DataType) - 类 中的静态方法org.apache.spark.sql.hive.execution.HiveScriptIOSchema
-
- toInspector(Expression) - 类 中的静态方法org.apache.spark.sql.hive.execution.HiveScriptIOSchema
-
- toInt() - 类 中的方法org.apache.spark.sql.types.Decimal
-
- toInt() - 类 中的方法org.apache.spark.storage.StorageLevel
-
- toIntArray() - 类 中的方法org.apache.spark.sql.vectorized.ColumnarArray
-
- toIterable() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- toIterator() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- toJavaBigDecimal() - 类 中的方法org.apache.spark.sql.types.Decimal
-
- toJavaBigInteger() - 类 中的方法org.apache.spark.sql.types.Decimal
-
- toJavaDStream() - 类 中的方法org.apache.spark.streaming.api.java.JavaPairDStream
-
Convert to a JavaDStream
- toJavaDStream() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- toJavaDStream() - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- toJavaRDD() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- toJavaRDD() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- toJavaRDD() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- toJavaRDD() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- toJavaRDD() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- toJavaRDD() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- toJavaRDD() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- toJavaRDD() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- toJavaRDD() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- toJavaRDD() - 类 中的方法org.apache.spark.rdd.RDD
-
- toJavaRDD() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- toJavaRDD() - 类 中的方法org.apache.spark.sql.Dataset
-
Returns the content of the Dataset as a JavaRDD of Ts.
- toJson(Matrix) - 类 中的静态方法org.apache.spark.ml.linalg.JsonMatrixConverter
-
Coverts the Matrix to a JSON string.
- toJson(Vector) - 类 中的静态方法org.apache.spark.ml.linalg.JsonVectorConverter
-
Coverts the vector to a JSON string.
- toJson() - 类 中的方法org.apache.spark.mllib.linalg.DenseVector
-
- toJson() - 类 中的方法org.apache.spark.mllib.linalg.SparseVector
-
- toJson() - 接口 中的方法org.apache.spark.mllib.linalg.Vector
-
Converts the vector to a JSON string.
- toJSON() - 类 中的方法org.apache.spark.sql.Dataset
-
Returns the content of the Dataset as a Dataset of JSON strings.
- toJSON() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- toJSON() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- toJSON() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- toJSON() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- Tokenizer - org.apache.spark.ml.feature中的类
-
A tokenizer that converts the input string to lowercase and then splits it by white spaces.
- Tokenizer(String) - 类 的构造器org.apache.spark.ml.feature.Tokenizer
-
- Tokenizer() - 类 的构造器org.apache.spark.ml.feature.Tokenizer
-
- tokens() - 类 中的方法org.apache.spark.scheduler.cluster.CoarseGrainedClusterMessages.UpdateDelegationTokens
-
- tol() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- tol() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- tol() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- tol() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- tol() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- tol() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixture
-
- tol() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- tol() - 类 中的静态方法org.apache.spark.ml.clustering.KMeans
-
- tol() - 类 中的静态方法org.apache.spark.ml.clustering.KMeansModel
-
- tol() - 接口 中的方法org.apache.spark.ml.param.shared.HasTol
-
Param for the convergence tolerance for iterative algorithms (>= 0).
- tol() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- tol() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- tol() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- tol() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- tol() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- tol() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- toList() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- toLocal() - 类 中的方法org.apache.spark.ml.clustering.DistributedLDAModel
-
Convert this distributed model to a local representation.
- toLocal() - 类 中的方法org.apache.spark.mllib.clustering.DistributedLDAModel
-
Convert model to a local model.
- toLocalIterator() - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- toLocalIterator() - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- toLocalIterator() - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- toLocalIterator() - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Return an iterator that contains all of the elements in this RDD.
- toLocalIterator() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- toLocalIterator() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- toLocalIterator() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- toLocalIterator() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- toLocalIterator() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- toLocalIterator() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- toLocalIterator() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- toLocalIterator() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- toLocalIterator() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- toLocalIterator() - 类 中的方法org.apache.spark.rdd.RDD
-
Return an iterator that contains all of the elements in this RDD.
- toLocalIterator() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- toLocalIterator() - 类 中的方法org.apache.spark.sql.Dataset
-
Returns an iterator that contains all rows in this Dataset.
- toLocalMatrix() - 类 中的方法org.apache.spark.mllib.linalg.distributed.BlockMatrix
-
Collect the distributed matrix on the driver as a DenseMatrix.
- toLong() - 类 中的方法org.apache.spark.sql.types.Decimal
-
- toLongArray() - 类 中的方法org.apache.spark.sql.vectorized.ColumnarArray
-
- toLowercase() - 类 中的方法org.apache.spark.ml.feature.RegexTokenizer
-
Indicates whether to convert all characters to lowercase before tokenizing.
- toMap(Predef.$less$colon$less<A, Tuple2<T, U>>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- toMetadata(Metadata) - 类 中的方法org.apache.spark.ml.attribute.Attribute
-
Converts to ML metadata with some existing metadata.
- toMetadata() - 类 中的方法org.apache.spark.ml.attribute.Attribute
-
Converts to ML metadata
- toMetadata(Metadata) - 类 中的方法org.apache.spark.ml.attribute.AttributeGroup
-
Converts to ML metadata with some existing metadata.
- toMetadata() - 类 中的方法org.apache.spark.ml.attribute.AttributeGroup
-
Converts to ML metadata
- toMetadata(Metadata) - 类 中的静态方法org.apache.spark.ml.attribute.BinaryAttribute
-
- toMetadata() - 类 中的静态方法org.apache.spark.ml.attribute.BinaryAttribute
-
- toMetadata(Metadata) - 类 中的静态方法org.apache.spark.ml.attribute.NominalAttribute
-
- toMetadata() - 类 中的静态方法org.apache.spark.ml.attribute.NominalAttribute
-
- toMetadata(Metadata) - 类 中的静态方法org.apache.spark.ml.attribute.NumericAttribute
-
- toMetadata() - 类 中的静态方法org.apache.spark.ml.attribute.NumericAttribute
-
- toMetadata(Metadata) - 类 中的静态方法org.apache.spark.ml.attribute.UnresolvedAttribute
-
- toMetadata() - 类 中的静态方法org.apache.spark.ml.attribute.UnresolvedAttribute
-
- toNetty() - 类 中的方法org.apache.spark.storage.DiskBlockData
-
Returns a Netty-friendly wrapper for the block's data.
- toNetty() - 类 中的方法org.apache.spark.storage.EncryptedBlockData
-
- toNumber(String, Function1<String, T>, String, String) - 类 中的静态方法org.apache.spark.internal.config.ConfigHelpers
-
- toOld() - 接口 中的方法org.apache.spark.ml.tree.Split
-
Convert to old Split format
- tooltip(String, String) - 类 中的静态方法org.apache.spark.ui.UIUtils
-
- ToolTips - org.apache.spark.ui中的类
-
- ToolTips() - 类 的构造器org.apache.spark.ui.ToolTips
-
- top(int, Comparator<T>) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- top(int) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- top(int, Comparator<T>) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- top(int) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- top(int, Comparator<T>) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- top(int) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- top(int, Comparator<T>) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Returns the top k (largest) elements from this RDD as defined by
the specified Comparator[T] and maintains the order.
- top(int) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Returns the top k (largest) elements from this RDD using the
natural ordering for T and maintains the order.
- top(int, Ordering<T>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- top(int, Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- top(int, Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- top(int, Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- top(int, Ordering<T>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- top(int, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- top(int, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- top(int, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- top(int, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- top(int, Ordering<T>) - 类 中的方法org.apache.spark.rdd.RDD
-
Returns the top k (largest) elements from this RDD as defined by the specified
implicit Ordering[T] and maintains the ordering.
- top(int, Ordering<T>) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- toPairDStreamFunctions(DStream<Tuple2<K, V>>, ClassTag<K>, ClassTag<V>, Ordering<K>) - 类 中的静态方法org.apache.spark.streaming.dstream.DStream
-
- topByKey(int, Ordering<V>) - 类 中的方法org.apache.spark.mllib.rdd.MLPairRDDFunctions
-
Returns the top k (largest) elements for each key from this RDD as defined by the specified
implicit Ordering[T].
- topDocumentsPerTopic(int) - 类 中的方法org.apache.spark.mllib.clustering.DistributedLDAModel
-
Return the top documents for each topic
- topicAssignments() - 类 中的方法org.apache.spark.mllib.clustering.DistributedLDAModel
-
Return the top topic for each (doc, term) pair.
- topicConcentration() - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- topicConcentration() - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- topicConcentration() - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- topicConcentration() - 类 中的方法org.apache.spark.mllib.clustering.DistributedLDAModel
-
- topicConcentration() - 类 中的方法org.apache.spark.mllib.clustering.LDAModel
-
Concentration parameter (commonly named "beta" or "eta") for the prior placed on topics'
distributions over terms.
- topicConcentration() - 类 中的方法org.apache.spark.mllib.clustering.LocalLDAModel
-
- topicDistribution(Vector) - 类 中的方法org.apache.spark.mllib.clustering.LocalLDAModel
-
Predicts the topic mixture distribution for a document (often called "theta" in the
literature).
- topicDistributionCol() - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- topicDistributionCol() - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- topicDistributionCol() - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- topicDistributions() - 类 中的方法org.apache.spark.mllib.clustering.DistributedLDAModel
-
For each document in the training set, return the distribution over topics for that document
("theta_doc").
- topicDistributions(RDD<Tuple2<Object, Vector>>) - 类 中的方法org.apache.spark.mllib.clustering.LocalLDAModel
-
Predicts the topic mixture distribution for each document (often called "theta" in the
literature).
- topicDistributions(JavaPairRDD<Long, Vector>) - 类 中的方法org.apache.spark.mllib.clustering.LocalLDAModel
-
Java-friendly version of topicDistributions
- topics() - 类 中的方法org.apache.spark.mllib.clustering.LocalLDAModel
-
- topicsMatrix() - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- topicsMatrix() - 类 中的方法org.apache.spark.ml.clustering.LDAModel
-
Inferred topics, where each topic is represented by a distribution over terms.
- topicsMatrix() - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- topicsMatrix() - 类 中的方法org.apache.spark.mllib.clustering.DistributedLDAModel
-
Inferred topics, where each topic is represented by a distribution over terms.
- topicsMatrix() - 类 中的方法org.apache.spark.mllib.clustering.LDAModel
-
Inferred topics, where each topic is represented by a distribution over terms.
- topicsMatrix() - 类 中的方法org.apache.spark.mllib.clustering.LocalLDAModel
-
- topK(Iterator<Tuple2<String, Object>>, int) - 类 中的静态方法org.apache.spark.streaming.util.RawTextHelper
-
Gets the top k words in terms of word counts.
- toPMML(String) - 类 中的静态方法org.apache.spark.mllib.classification.LogisticRegressionModel
-
- toPMML(SparkContext, String) - 类 中的静态方法org.apache.spark.mllib.classification.LogisticRegressionModel
-
- toPMML(OutputStream) - 类 中的静态方法org.apache.spark.mllib.classification.LogisticRegressionModel
-
- toPMML() - 类 中的静态方法org.apache.spark.mllib.classification.LogisticRegressionModel
-
- toPMML(String) - 类 中的静态方法org.apache.spark.mllib.classification.SVMModel
-
- toPMML(SparkContext, String) - 类 中的静态方法org.apache.spark.mllib.classification.SVMModel
-
- toPMML(OutputStream) - 类 中的静态方法org.apache.spark.mllib.classification.SVMModel
-
- toPMML() - 类 中的静态方法org.apache.spark.mllib.classification.SVMModel
-
- toPMML(String) - 类 中的静态方法org.apache.spark.mllib.clustering.KMeansModel
-
- toPMML(SparkContext, String) - 类 中的静态方法org.apache.spark.mllib.clustering.KMeansModel
-
- toPMML(OutputStream) - 类 中的静态方法org.apache.spark.mllib.clustering.KMeansModel
-
- toPMML() - 类 中的静态方法org.apache.spark.mllib.clustering.KMeansModel
-
- toPMML(StreamResult) - 接口 中的方法org.apache.spark.mllib.pmml.PMMLExportable
-
Export the model to the stream result in PMML format
- toPMML(String) - 接口 中的方法org.apache.spark.mllib.pmml.PMMLExportable
-
Export the model to a local file in PMML format
- toPMML(SparkContext, String) - 接口 中的方法org.apache.spark.mllib.pmml.PMMLExportable
-
Export the model to a directory on a distributed file system in PMML format
- toPMML(OutputStream) - 接口 中的方法org.apache.spark.mllib.pmml.PMMLExportable
-
Export the model to the OutputStream in PMML format
- toPMML() - 接口 中的方法org.apache.spark.mllib.pmml.PMMLExportable
-
Export the model to a String in PMML format
- toPMML(String) - 类 中的静态方法org.apache.spark.mllib.regression.LassoModel
-
- toPMML(SparkContext, String) - 类 中的静态方法org.apache.spark.mllib.regression.LassoModel
-
- toPMML(OutputStream) - 类 中的静态方法org.apache.spark.mllib.regression.LassoModel
-
- toPMML() - 类 中的静态方法org.apache.spark.mllib.regression.LassoModel
-
- toPMML(String) - 类 中的静态方法org.apache.spark.mllib.regression.LinearRegressionModel
-
- toPMML(SparkContext, String) - 类 中的静态方法org.apache.spark.mllib.regression.LinearRegressionModel
-
- toPMML(OutputStream) - 类 中的静态方法org.apache.spark.mllib.regression.LinearRegressionModel
-
- toPMML() - 类 中的静态方法org.apache.spark.mllib.regression.LinearRegressionModel
-
- toPMML(String) - 类 中的静态方法org.apache.spark.mllib.regression.RidgeRegressionModel
-
- toPMML(SparkContext, String) - 类 中的静态方法org.apache.spark.mllib.regression.RidgeRegressionModel
-
- toPMML(OutputStream) - 类 中的静态方法org.apache.spark.mllib.regression.RidgeRegressionModel
-
- toPMML() - 类 中的静态方法org.apache.spark.mllib.regression.RidgeRegressionModel
-
- topNode() - 类 中的方法org.apache.spark.mllib.tree.model.DecisionTreeModel
-
- topologyFile() - 类 中的方法org.apache.spark.storage.FileBasedTopologyMapper
-
- topologyInfo() - 类 中的方法org.apache.spark.storage.BlockManagerId
-
- topologyMap() - 类 中的方法org.apache.spark.storage.FileBasedTopologyMapper
-
- TopologyMapper - org.apache.spark.storage中的类
-
::DeveloperApi::
TopologyMapper provides topology information for a given host
param: conf SparkConf to get required properties, if needed
- TopologyMapper(SparkConf) - 类 的构造器org.apache.spark.storage.TopologyMapper
-
- toPredict() - 类 中的方法org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.PredictData
-
- topTopicsPerDocument(int) - 类 中的方法org.apache.spark.mllib.clustering.DistributedLDAModel
-
For each document, return the top k weighted topics for that document and their weights.
- toRadians(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
- toRadians(String) - 类 中的静态方法org.apache.spark.sql.functions
-
- toRDD(JavaDoubleRDD) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- toRDD(JavaPairRDD<K, V>) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- toRDD(JavaRDD<T>) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- toRowMatrix() - 类 中的方法org.apache.spark.mllib.linalg.distributed.CoordinateMatrix
-
Converts to RowMatrix, dropping row indices after grouping by row index.
- toRowMatrix() - 类 中的方法org.apache.spark.mllib.linalg.distributed.IndexedRowMatrix
-
Drops row indices and converts this matrix to a
RowMatrix.
- toScalaBigInt() - 类 中的方法org.apache.spark.sql.types.Decimal
-
- toSeq() - 类 中的方法org.apache.spark.ml.param.ParamMap
-
Converts this param map to a sequence of param pairs.
- toSeq() - 接口 中的方法org.apache.spark.sql.Row
-
Return a Scala Seq representing the row.
- toSeq() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- toSet() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- toShort() - 类 中的方法org.apache.spark.sql.types.Decimal
-
- toShortArray() - 类 中的方法org.apache.spark.sql.vectorized.ColumnarArray
-
- toSparkContext(JavaSparkContext) - 类 中的静态方法org.apache.spark.api.java.JavaSparkContext
-
- toSparse() - 类 中的静态方法org.apache.spark.ml.linalg.DenseMatrix
-
- toSparse() - 类 中的静态方法org.apache.spark.ml.linalg.DenseVector
-
- toSparse() - 接口 中的方法org.apache.spark.ml.linalg.Matrix
-
Converts this matrix to a sparse matrix while maintaining the layout of the current matrix.
- toSparse() - 类 中的静态方法org.apache.spark.ml.linalg.SparseMatrix
-
- toSparse() - 类 中的静态方法org.apache.spark.ml.linalg.SparseVector
-
- toSparse() - 接口 中的方法org.apache.spark.ml.linalg.Vector
-
Converts this vector to a sparse vector with all explicit zeros removed.
- toSparse() - 类 中的方法org.apache.spark.mllib.linalg.DenseMatrix
-
Generate a SparseMatrix from the given DenseMatrix.
- toSparse() - 类 中的静态方法org.apache.spark.mllib.linalg.DenseVector
-
- toSparse() - 类 中的静态方法org.apache.spark.mllib.linalg.SparseVector
-
- toSparse() - 接口 中的方法org.apache.spark.mllib.linalg.Vector
-
Converts this vector to a sparse vector with all explicit zeros removed.
- toSparseColMajor() - 类 中的静态方法org.apache.spark.ml.linalg.DenseMatrix
-
- toSparseColMajor() - 接口 中的方法org.apache.spark.ml.linalg.Matrix
-
Converts this matrix to a sparse matrix in column major order.
- toSparseColMajor() - 类 中的静态方法org.apache.spark.ml.linalg.SparseMatrix
-
- toSparseMatrix(boolean) - 接口 中的方法org.apache.spark.ml.linalg.Matrix
-
Converts this matrix to a sparse matrix.
- toSparseRowMajor() - 类 中的静态方法org.apache.spark.ml.linalg.DenseMatrix
-
- toSparseRowMajor() - 接口 中的方法org.apache.spark.ml.linalg.Matrix
-
Converts this matrix to a sparse matrix in row major order.
- toSparseRowMajor() - 类 中的静态方法org.apache.spark.ml.linalg.SparseMatrix
-
- toSparseWithSize(int) - 接口 中的方法org.apache.spark.ml.linalg.Vector
-
Converts this vector to a sparse vector with all explicit zeros removed when the size is known.
- toSparseWithSize(int) - 接口 中的方法org.apache.spark.mllib.linalg.Vector
-
Converts this vector to a sparse vector with all explicit zeros removed when the size is known.
- toSplit() - 类 中的方法org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.SplitData
-
- toSplitInfo(Class<?>, String, InputSplit) - 类 中的静态方法org.apache.spark.scheduler.SplitInfo
-
- toSplitInfo(Class<?>, String, InputSplit) - 类 中的静态方法org.apache.spark.scheduler.SplitInfo
-
- toStream() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- toString() - 类 中的方法org.apache.spark.Accumulable
-
已过时。
- toString() - 类 中的静态方法org.apache.spark.Accumulator
-
已过时。
- toString() - 类 中的方法org.apache.spark.api.java.JavaRDD
-
- toString() - 类 中的方法org.apache.spark.api.java.Optional
-
- toString() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- toString() - 类 中的方法org.apache.spark.broadcast.Broadcast
-
- toString() - 类 中的方法org.apache.spark.graphx.EdgeDirection
-
- toString() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- toString() - 类 中的方法org.apache.spark.graphx.EdgeTriplet
-
- toString() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- toString() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- toString() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- toString() - 类 中的方法org.apache.spark.ml.attribute.Attribute
-
- toString() - 类 中的方法org.apache.spark.ml.attribute.AttributeGroup
-
- toString() - 类 中的静态方法org.apache.spark.ml.attribute.BinaryAttribute
-
- toString() - 类 中的静态方法org.apache.spark.ml.attribute.NominalAttribute
-
- toString() - 类 中的静态方法org.apache.spark.ml.attribute.NumericAttribute
-
- toString() - 类 中的静态方法org.apache.spark.ml.attribute.UnresolvedAttribute
-
- toString() - 类 中的方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- toString() - 类 中的方法org.apache.spark.ml.classification.GBTClassificationModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- toString() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- toString() - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- toString() - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- toString() - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- toString() - 类 中的方法org.apache.spark.ml.classification.NaiveBayesModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.classification.OneVsRest
-
- toString() - 类 中的静态方法org.apache.spark.ml.classification.OneVsRestModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- toString() - 类 中的方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- toString() - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeans
-
- toString() - 类 中的静态方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixture
-
- toString() - 类 中的静态方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.clustering.KMeans
-
- toString() - 类 中的静态方法org.apache.spark.ml.clustering.KMeansModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.clustering.LDA
-
- toString() - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
-
- toString() - 类 中的静态方法org.apache.spark.ml.evaluation.ClusteringEvaluator
-
- toString() - 类 中的静态方法org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
-
- toString() - 类 中的静态方法org.apache.spark.ml.evaluation.RegressionEvaluator
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.Binarizer
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.Bucketizer
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelector
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.ColumnPruner
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizer
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.CountVectorizerModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.DCT
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.ElementwiseProduct
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.FeatureHasher
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.HashingTF
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.IDF
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.IDFModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.Imputer
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.ImputerModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.IndexToString
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.Interaction
-
- toString() - 类 中的方法org.apache.spark.ml.feature.LabeledPoint
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.MaxAbsScaler
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.MaxAbsScalerModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSH
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSHModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScaler
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.MinMaxScalerModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.NGram
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.Normalizer
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoder
-
已过时。
- toString() - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderEstimator
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.PCA
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.PCAModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.PolynomialExpansion
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.QuantileDiscretizer
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.RegexTokenizer
-
- toString() - 类 中的方法org.apache.spark.ml.feature.RFormula
-
- toString() - 类 中的方法org.apache.spark.ml.feature.RFormulaModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.SQLTransformer
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.StandardScaler
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.StandardScalerModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.StopWordsRemover
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.StringIndexer
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.StringIndexerModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.Tokenizer
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.VectorAssembler
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.VectorAttributeRewriter
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexer
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.VectorIndexerModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.VectorSizeHint
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.VectorSlicer
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.Word2Vec
-
- toString() - 类 中的静态方法org.apache.spark.ml.feature.Word2VecModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowth
-
- toString() - 类 中的静态方法org.apache.spark.ml.fpm.FPGrowthModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.linalg.DenseMatrix
-
- toString(int, int) - 类 中的静态方法org.apache.spark.ml.linalg.DenseMatrix
-
- toString() - 类 中的方法org.apache.spark.ml.linalg.DenseVector
-
- toString() - 接口 中的方法org.apache.spark.ml.linalg.Matrix
-
A human readable representation of the matrix
- toString(int, int) - 接口 中的方法org.apache.spark.ml.linalg.Matrix
-
A human readable representation of the matrix with maximum lines and width
- toString() - 类 中的静态方法org.apache.spark.ml.linalg.SparseMatrix
-
- toString(int, int) - 类 中的静态方法org.apache.spark.ml.linalg.SparseMatrix
-
- toString() - 类 中的方法org.apache.spark.ml.linalg.SparseVector
-
- toString() - 类 中的静态方法org.apache.spark.ml.param.DoubleParam
-
- toString() - 类 中的静态方法org.apache.spark.ml.param.FloatParam
-
- toString() - 类 中的方法org.apache.spark.ml.param.Param
-
- toString() - 类 中的方法org.apache.spark.ml.param.ParamMap
-
- toString() - 类 中的静态方法org.apache.spark.ml.Pipeline
-
- toString() - 类 中的静态方法org.apache.spark.ml.PipelineModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
- toString() - 类 中的静态方法org.apache.spark.ml.recommendation.ALSModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- toString() - 类 中的静态方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- toString() - 类 中的方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- toString() - 类 中的方法org.apache.spark.ml.regression.GBTRegressionModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- toString() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- toString() - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- toString() - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegressionTrainingSummary
-
- toString() - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegression
-
- toString() - 类 中的静态方法org.apache.spark.ml.regression.IsotonicRegressionModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- toString() - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- toString() - 类 中的方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- toString() - 类 中的方法org.apache.spark.ml.tree.InternalNode
-
- toString() - 类 中的方法org.apache.spark.ml.tree.LeafNode
-
- toString() - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidator
-
- toString() - 类 中的静态方法org.apache.spark.ml.tuning.CrossValidatorModel
-
- toString() - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplit
-
- toString() - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplitModel
-
- toString() - 接口 中的方法org.apache.spark.ml.util.Identifiable
-
- toString() - 类 中的方法org.apache.spark.mllib.classification.LogisticRegressionModel
-
- toString() - 类 中的方法org.apache.spark.mllib.classification.SVMModel
-
- toString() - 类 中的方法org.apache.spark.mllib.fpm.AssociationRules.Rule
-
- toString() - 类 中的方法org.apache.spark.mllib.fpm.FPGrowth.FreqItemset
-
- toString() - 类 中的静态方法org.apache.spark.mllib.linalg.DenseMatrix
-
- toString(int, int) - 类 中的静态方法org.apache.spark.mllib.linalg.DenseMatrix
-
- toString() - 类 中的方法org.apache.spark.mllib.linalg.DenseVector
-
- toString() - 接口 中的方法org.apache.spark.mllib.linalg.Matrix
-
A human readable representation of the matrix
- toString(int, int) - 接口 中的方法org.apache.spark.mllib.linalg.Matrix
-
A human readable representation of the matrix with maximum lines and width
- toString() - 类 中的静态方法org.apache.spark.mllib.linalg.SparseMatrix
-
- toString(int, int) - 类 中的静态方法org.apache.spark.mllib.linalg.SparseMatrix
-
- toString() - 类 中的方法org.apache.spark.mllib.linalg.SparseVector
-
- toString() - 类 中的方法org.apache.spark.mllib.regression.GeneralizedLinearModel
-
Print a summary of the model.
- toString() - 类 中的方法org.apache.spark.mllib.regression.LabeledPoint
-
- toString() - 类 中的静态方法org.apache.spark.mllib.regression.LassoModel
-
- toString() - 类 中的静态方法org.apache.spark.mllib.regression.LinearRegressionModel
-
- toString() - 类 中的静态方法org.apache.spark.mllib.regression.RidgeRegressionModel
-
- toString() - 类 中的方法org.apache.spark.mllib.stat.test.BinarySample
-
- toString() - 类 中的方法org.apache.spark.mllib.stat.test.ChiSqTestResult
-
- toString() - 类 中的方法org.apache.spark.mllib.stat.test.KolmogorovSmirnovTestResult
-
- toString() - 接口 中的方法org.apache.spark.mllib.stat.test.TestResult
-
String explaining the hypothesis test result.
- toString() - 类 中的静态方法org.apache.spark.mllib.tree.configuration.Algo
-
- toString() - 类 中的静态方法org.apache.spark.mllib.tree.configuration.EnsembleCombiningStrategy
-
- toString() - 类 中的静态方法org.apache.spark.mllib.tree.configuration.FeatureType
-
- toString() - 类 中的静态方法org.apache.spark.mllib.tree.configuration.QuantileStrategy
-
- toString() - 类 中的方法org.apache.spark.mllib.tree.model.DecisionTreeModel
-
Print a summary of the model.
- toString() - 类 中的静态方法org.apache.spark.mllib.tree.model.GradientBoostedTreesModel
-
- toString() - 类 中的方法org.apache.spark.mllib.tree.model.InformationGainStats
-
- toString() - 类 中的方法org.apache.spark.mllib.tree.model.Node
-
- toString() - 类 中的方法org.apache.spark.mllib.tree.model.Predict
-
- toString() - 类 中的静态方法org.apache.spark.mllib.tree.model.RandomForestModel
-
- toString() - 类 中的方法org.apache.spark.mllib.tree.model.Split
-
- toString() - 类 中的方法org.apache.spark.partial.BoundedDouble
-
- toString() - 类 中的方法org.apache.spark.partial.PartialResult
-
- toString() - 类 中的静态方法org.apache.spark.rdd.CheckpointState
-
- toString() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- toString() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- toString() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- toString() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- toString() - 类 中的方法org.apache.spark.rdd.RDD
-
- toString() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- toString() - 类 中的静态方法org.apache.spark.scheduler.ExecutorKilled
-
- toString() - 类 中的方法org.apache.spark.scheduler.InputFormatInfo
-
- toString() - 类 中的静态方法org.apache.spark.scheduler.LossReasonPending
-
- toString() - 类 中的静态方法org.apache.spark.scheduler.SchedulingMode
-
- toString() - 类 中的方法org.apache.spark.scheduler.SplitInfo
-
- toString() - 类 中的静态方法org.apache.spark.scheduler.TaskLocality
-
- toString() - 类 中的方法org.apache.spark.SerializableWritable
-
- toString() - 异常错误 中的静态方法org.apache.spark.sql.AnalysisException
-
- toString() - 类 中的方法org.apache.spark.sql.catalog.Column
-
- toString() - 类 中的方法org.apache.spark.sql.catalog.Database
-
- toString() - 类 中的方法org.apache.spark.sql.catalog.Function
-
- toString() - 类 中的方法org.apache.spark.sql.catalog.Table
-
- toString() - 类 中的方法org.apache.spark.sql.Column
-
- toString() - 类 中的方法org.apache.spark.sql.Dataset
-
- toString() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- toString() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- toString() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- toString() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- toString() - 类 中的方法org.apache.spark.sql.hive.orc.OrcFileFormat
-
- toString() - 类 中的方法org.apache.spark.sql.KeyValueGroupedDataset
-
- toString() - 类 中的方法org.apache.spark.sql.RelationalGroupedDataset
-
- toString() - 接口 中的方法org.apache.spark.sql.Row
-
- toString() - 类 中的方法org.apache.spark.sql.sources.In
-
- toString() - 类 中的方法org.apache.spark.sql.sources.v2.reader.streaming.Offset
-
- toString() - 类 中的方法org.apache.spark.sql.streaming.SinkProgress
-
- toString() - 类 中的方法org.apache.spark.sql.streaming.SourceProgress
-
- toString() - 类 中的方法org.apache.spark.sql.streaming.StateOperatorProgress
-
- toString() - 异常错误 中的方法org.apache.spark.sql.streaming.StreamingQueryException
-
- toString() - 类 中的方法org.apache.spark.sql.streaming.StreamingQueryProgress
-
- toString() - 类 中的方法org.apache.spark.sql.streaming.StreamingQueryStatus
-
- toString() - 类 中的方法org.apache.spark.sql.types.Decimal
-
- toString() - 类 中的方法org.apache.spark.sql.types.DecimalType
-
- toString() - 类 中的方法org.apache.spark.sql.types.Metadata
-
- toString() - 类 中的方法org.apache.spark.sql.types.StructField
-
- toString() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- toString() - 类 中的方法org.apache.spark.storage.BlockId
-
- toString() - 类 中的方法org.apache.spark.storage.BlockManagerId
-
- toString() - 类 中的静态方法org.apache.spark.storage.BroadcastBlockId
-
- toString() - 类 中的静态方法org.apache.spark.storage.RDDBlockId
-
- toString() - 类 中的方法org.apache.spark.storage.RDDInfo
-
- toString() - 类 中的静态方法org.apache.spark.storage.ShuffleBlockId
-
- toString() - 类 中的静态方法org.apache.spark.storage.ShuffleDataBlockId
-
- toString() - 类 中的静态方法org.apache.spark.storage.ShuffleIndexBlockId
-
- toString() - 类 中的方法org.apache.spark.storage.StorageLevel
-
- toString() - 类 中的静态方法org.apache.spark.storage.StreamBlockId
-
- toString() - 类 中的静态方法org.apache.spark.storage.TaskResultBlockId
-
- toString() - 类 中的方法org.apache.spark.streaming.Duration
-
- toString() - 类 中的静态方法org.apache.spark.streaming.scheduler.ReceiverState
-
- toString() - 类 中的方法org.apache.spark.streaming.State
-
- toString() - 类 中的方法org.apache.spark.streaming.Time
-
- toString() - 类 中的静态方法org.apache.spark.TaskState
-
- toString() - 类 中的方法org.apache.spark.util.AccumulatorV2
-
- toString() - 类 中的方法org.apache.spark.util.MutablePair
-
- toString() - 类 中的方法org.apache.spark.util.StatCounter
-
- toStructField(Metadata) - 类 中的方法org.apache.spark.ml.attribute.Attribute
-
Converts to a StructField with some existing metadata.
- toStructField() - 类 中的方法org.apache.spark.ml.attribute.Attribute
-
Converts to a StructField.
- toStructField(Metadata) - 类 中的方法org.apache.spark.ml.attribute.AttributeGroup
-
Converts to a StructField with some existing metadata.
- toStructField() - 类 中的方法org.apache.spark.ml.attribute.AttributeGroup
-
Converts to a StructField.
- toStructField(Metadata) - 类 中的静态方法org.apache.spark.ml.attribute.BinaryAttribute
-
- toStructField() - 类 中的静态方法org.apache.spark.ml.attribute.BinaryAttribute
-
- toStructField(Metadata) - 类 中的静态方法org.apache.spark.ml.attribute.NominalAttribute
-
- toStructField() - 类 中的静态方法org.apache.spark.ml.attribute.NominalAttribute
-
- toStructField(Metadata) - 类 中的静态方法org.apache.spark.ml.attribute.NumericAttribute
-
- toStructField() - 类 中的静态方法org.apache.spark.ml.attribute.NumericAttribute
-
- toStructField(Metadata) - 类 中的静态方法org.apache.spark.ml.attribute.UnresolvedAttribute
-
- toStructField() - 类 中的静态方法org.apache.spark.ml.attribute.UnresolvedAttribute
-
- totalBlocksFetched() - 类 中的方法org.apache.spark.status.api.v1.ShuffleReadMetricDistributions
-
- totalBytesRead(ShuffleReadMetrics) - 类 中的静态方法org.apache.spark.ui.jobs.ApiHelper
-
- totalCores() - 类 中的方法org.apache.spark.scheduler.cluster.ExecutorInfo
-
- totalCores() - 类 中的方法org.apache.spark.status.api.v1.ExecutorSummary
-
- totalCores() - 类 中的方法org.apache.spark.status.LiveExecutor
-
- totalCount() - 类 中的方法org.apache.spark.util.sketch.CountMinSketch
-
- totalDelay() - 类 中的方法org.apache.spark.status.api.v1.streaming.BatchInfo
-
- totalDelay() - 类 中的方法org.apache.spark.streaming.scheduler.BatchInfo
-
Time taken for all the jobs of this batch to finish processing from the time they
were submitted.
- totalDiskSize() - 类 中的方法org.apache.spark.ui.storage.ExecutorStreamSummary
-
- totalDuration() - 类 中的方法org.apache.spark.status.api.v1.ExecutorSummary
-
- totalDuration() - 类 中的方法org.apache.spark.status.LiveExecutor
-
- totalGCTime() - 类 中的方法org.apache.spark.status.api.v1.ExecutorSummary
-
- totalGcTime() - 类 中的方法org.apache.spark.status.LiveExecutor
-
- totalInputBytes() - 类 中的方法org.apache.spark.status.api.v1.ExecutorSummary
-
- totalInputBytes() - 类 中的方法org.apache.spark.status.LiveExecutor
-
- totalIterations() - 接口 中的方法org.apache.spark.ml.classification.LogisticRegressionTrainingSummary
-
Number of training iterations.
- totalIterations() - 类 中的方法org.apache.spark.ml.regression.LinearRegressionTrainingSummary
-
Number of training iterations until termination
This value is only available when using the "l-bfgs" solver.
- totalMemSize() - 类 中的方法org.apache.spark.ui.storage.ExecutorStreamSummary
-
- totalNumNodes() - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- totalNumNodes() - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- totalNumNodes() - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- totalNumNodes() - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- totalNumNodes() - 类 中的静态方法org.apache.spark.mllib.tree.model.GradientBoostedTreesModel
-
- totalNumNodes() - 类 中的静态方法org.apache.spark.mllib.tree.model.RandomForestModel
-
- totalOffHeap() - 类 中的方法org.apache.spark.status.LiveExecutor
-
- totalOffHeapStorageMemory() - 类 中的方法org.apache.spark.status.api.v1.MemoryMetrics
-
- totalOnHeap() - 类 中的方法org.apache.spark.status.LiveExecutor
-
- totalOnHeapStorageMemory() - 类 中的方法org.apache.spark.status.api.v1.MemoryMetrics
-
- totalShuffleRead() - 类 中的方法org.apache.spark.status.api.v1.ExecutorSummary
-
- totalShuffleRead() - 类 中的方法org.apache.spark.status.LiveExecutor
-
- totalShuffleWrite() - 类 中的方法org.apache.spark.status.api.v1.ExecutorSummary
-
- totalShuffleWrite() - 类 中的方法org.apache.spark.status.LiveExecutor
-
- totalTasks() - 类 中的方法org.apache.spark.status.api.v1.ExecutorSummary
-
- totalTasks() - 类 中的方法org.apache.spark.status.LiveExecutor
-
- toTraversable() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- toTuple() - 类 中的方法org.apache.spark.graphx.EdgeTriplet
-
- toUnscaledLong() - 类 中的方法org.apache.spark.sql.types.Decimal
-
- toVector() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- toVirtualHosts(Seq<String>) - 类 中的静态方法org.apache.spark.ui.JettyUtils
-
- train(RDD<ALS.Rating<ID>>, int, int, int, int, double, boolean, double, boolean, StorageLevel, StorageLevel, int, long, ClassTag<ID>, Ordering<ID>) - 类 中的静态方法org.apache.spark.ml.recommendation.ALS
-
:: DeveloperApi ::
Implementation of the ALS algorithm.
- train(RDD<LabeledPoint>, int, double, double, Vector) - 类 中的静态方法org.apache.spark.mllib.classification.LogisticRegressionWithSGD
-
Train a logistic regression model given an RDD of (label, features) pairs.
- train(RDD<LabeledPoint>, int, double, double) - 类 中的静态方法org.apache.spark.mllib.classification.LogisticRegressionWithSGD
-
Train a logistic regression model given an RDD of (label, features) pairs.
- train(RDD<LabeledPoint>, int, double) - 类 中的静态方法org.apache.spark.mllib.classification.LogisticRegressionWithSGD
-
Train a logistic regression model given an RDD of (label, features) pairs.
- train(RDD<LabeledPoint>, int) - 类 中的静态方法org.apache.spark.mllib.classification.LogisticRegressionWithSGD
-
Train a logistic regression model given an RDD of (label, features) pairs.
- train(RDD<LabeledPoint>) - 类 中的静态方法org.apache.spark.mllib.classification.NaiveBayes
-
Trains a Naive Bayes model given an RDD of (label, features) pairs.
- train(RDD<LabeledPoint>, double) - 类 中的静态方法org.apache.spark.mllib.classification.NaiveBayes
-
Trains a Naive Bayes model given an RDD of (label, features) pairs.
- train(RDD<LabeledPoint>, double, String) - 类 中的静态方法org.apache.spark.mllib.classification.NaiveBayes
-
Trains a Naive Bayes model given an RDD of (label, features) pairs.
- train(RDD<LabeledPoint>, int, double, double, double, Vector) - 类 中的静态方法org.apache.spark.mllib.classification.SVMWithSGD
-
Train a SVM model given an RDD of (label, features) pairs.
- train(RDD<LabeledPoint>, int, double, double, double) - 类 中的静态方法org.apache.spark.mllib.classification.SVMWithSGD
-
Train a SVM model given an RDD of (label, features) pairs.
- train(RDD<LabeledPoint>, int, double, double) - 类 中的静态方法org.apache.spark.mllib.classification.SVMWithSGD
-
Train a SVM model given an RDD of (label, features) pairs.
- train(RDD<LabeledPoint>, int) - 类 中的静态方法org.apache.spark.mllib.classification.SVMWithSGD
-
Train a SVM model given an RDD of (label, features) pairs.
- train(RDD<Vector>, int, int, String, long) - 类 中的静态方法org.apache.spark.mllib.clustering.KMeans
-
Trains a k-means model using the given set of parameters.
- train(RDD<Vector>, int, int, String) - 类 中的静态方法org.apache.spark.mllib.clustering.KMeans
-
Trains a k-means model using the given set of parameters.
- train(RDD<Vector>, int, int, int, String, long) - 类 中的静态方法org.apache.spark.mllib.clustering.KMeans
-
- train(RDD<Vector>, int, int, int, String) - 类 中的静态方法org.apache.spark.mllib.clustering.KMeans
-
- train(RDD<Vector>, int, int) - 类 中的静态方法org.apache.spark.mllib.clustering.KMeans
-
Trains a k-means model using specified parameters and the default values for unspecified.
- train(RDD<Vector>, int, int, int) - 类 中的静态方法org.apache.spark.mllib.clustering.KMeans
-
- train(RDD<Rating>, int, int, double, int, long) - 类 中的静态方法org.apache.spark.mllib.recommendation.ALS
-
Train a matrix factorization model given an RDD of ratings by users for a subset of products.
- train(RDD<Rating>, int, int, double, int) - 类 中的静态方法org.apache.spark.mllib.recommendation.ALS
-
Train a matrix factorization model given an RDD of ratings by users for a subset of products.
- train(RDD<Rating>, int, int, double) - 类 中的静态方法org.apache.spark.mllib.recommendation.ALS
-
Train a matrix factorization model given an RDD of ratings by users for a subset of products.
- train(RDD<Rating>, int, int) - 类 中的静态方法org.apache.spark.mllib.recommendation.ALS
-
Train a matrix factorization model given an RDD of ratings by users for a subset of products.
- train(RDD<LabeledPoint>, int, double, double, double, Vector) - 类 中的静态方法org.apache.spark.mllib.regression.LassoWithSGD
-
Train a Lasso model given an RDD of (label, features) pairs.
- train(RDD<LabeledPoint>, int, double, double, double) - 类 中的静态方法org.apache.spark.mllib.regression.LassoWithSGD
-
Train a Lasso model given an RDD of (label, features) pairs.
- train(RDD<LabeledPoint>, int, double, double) - 类 中的静态方法org.apache.spark.mllib.regression.LassoWithSGD
-
Train a Lasso model given an RDD of (label, features) pairs.
- train(RDD<LabeledPoint>, int) - 类 中的静态方法org.apache.spark.mllib.regression.LassoWithSGD
-
Train a Lasso model given an RDD of (label, features) pairs.
- train(RDD<LabeledPoint>, int, double, double, Vector) - 类 中的静态方法org.apache.spark.mllib.regression.LinearRegressionWithSGD
-
Train a Linear Regression model given an RDD of (label, features) pairs.
- train(RDD<LabeledPoint>, int, double, double) - 类 中的静态方法org.apache.spark.mllib.regression.LinearRegressionWithSGD
-
Train a LinearRegression model given an RDD of (label, features) pairs.
- train(RDD<LabeledPoint>, int, double) - 类 中的静态方法org.apache.spark.mllib.regression.LinearRegressionWithSGD
-
Train a LinearRegression model given an RDD of (label, features) pairs.
- train(RDD<LabeledPoint>, int) - 类 中的静态方法org.apache.spark.mllib.regression.LinearRegressionWithSGD
-
Train a LinearRegression model given an RDD of (label, features) pairs.
- train(RDD<LabeledPoint>, int, double, double, double, Vector) - 类 中的静态方法org.apache.spark.mllib.regression.RidgeRegressionWithSGD
-
Train a RidgeRegression model given an RDD of (label, features) pairs.
- train(RDD<LabeledPoint>, int, double, double, double) - 类 中的静态方法org.apache.spark.mllib.regression.RidgeRegressionWithSGD
-
Train a RidgeRegression model given an RDD of (label, features) pairs.
- train(RDD<LabeledPoint>, int, double, double) - 类 中的静态方法org.apache.spark.mllib.regression.RidgeRegressionWithSGD
-
Train a RidgeRegression model given an RDD of (label, features) pairs.
- train(RDD<LabeledPoint>, int) - 类 中的静态方法org.apache.spark.mllib.regression.RidgeRegressionWithSGD
-
Train a RidgeRegression model given an RDD of (label, features) pairs.
- train(RDD<LabeledPoint>, Strategy) - 类 中的静态方法org.apache.spark.mllib.tree.DecisionTree
-
Method to train a decision tree model.
- train(RDD<LabeledPoint>, Enumeration.Value, Impurity, int) - 类 中的静态方法org.apache.spark.mllib.tree.DecisionTree
-
Method to train a decision tree model.
- train(RDD<LabeledPoint>, Enumeration.Value, Impurity, int, int) - 类 中的静态方法org.apache.spark.mllib.tree.DecisionTree
-
Method to train a decision tree model.
- train(RDD<LabeledPoint>, Enumeration.Value, Impurity, int, int, int, Enumeration.Value, Map<Object, Object>) - 类 中的静态方法org.apache.spark.mllib.tree.DecisionTree
-
Method to train a decision tree model.
- train(RDD<LabeledPoint>, BoostingStrategy) - 类 中的静态方法org.apache.spark.mllib.tree.GradientBoostedTrees
-
Method to train a gradient boosting model.
- train(JavaRDD<LabeledPoint>, BoostingStrategy) - 类 中的静态方法org.apache.spark.mllib.tree.GradientBoostedTrees
-
Java-friendly API for org.apache.spark.mllib.tree.GradientBoostedTrees.train
- trainClassifier(RDD<LabeledPoint>, int, Map<Object, Object>, String, int, int) - 类 中的静态方法org.apache.spark.mllib.tree.DecisionTree
-
Method to train a decision tree model for binary or multiclass classification.
- trainClassifier(JavaRDD<LabeledPoint>, int, Map<Integer, Integer>, String, int, int) - 类 中的静态方法org.apache.spark.mllib.tree.DecisionTree
-
Java-friendly API for org.apache.spark.mllib.tree.DecisionTree.trainClassifier
- trainClassifier(RDD<LabeledPoint>, Strategy, int, String, int) - 类 中的静态方法org.apache.spark.mllib.tree.RandomForest
-
Method to train a decision tree model for binary or multiclass classification.
- trainClassifier(RDD<LabeledPoint>, int, Map<Object, Object>, int, String, String, int, int, int) - 类 中的静态方法org.apache.spark.mllib.tree.RandomForest
-
Method to train a decision tree model for binary or multiclass classification.
- trainClassifier(JavaRDD<LabeledPoint>, int, Map<Integer, Integer>, int, String, String, int, int, int) - 类 中的静态方法org.apache.spark.mllib.tree.RandomForest
-
Java-friendly API for org.apache.spark.mllib.tree.RandomForest.trainClassifier
- trainImplicit(RDD<Rating>, int, int, double, int, double, long) - 类 中的静态方法org.apache.spark.mllib.recommendation.ALS
-
Train a matrix factorization model given an RDD of 'implicit preferences' given by users
to some products, in the form of (userID, productID, preference) pairs.
- trainImplicit(RDD<Rating>, int, int, double, int, double) - 类 中的静态方法org.apache.spark.mllib.recommendation.ALS
-
Train a matrix factorization model given an RDD of 'implicit preferences' of users for a
subset of products.
- trainImplicit(RDD<Rating>, int, int, double, double) - 类 中的静态方法org.apache.spark.mllib.recommendation.ALS
-
Train a matrix factorization model given an RDD of 'implicit preferences' of users for a
subset of products.
- trainImplicit(RDD<Rating>, int, int) - 类 中的静态方法org.apache.spark.mllib.recommendation.ALS
-
Train a matrix factorization model given an RDD of 'implicit preferences' of users for a
subset of products.
- trainingLogLikelihood() - 类 中的方法org.apache.spark.ml.clustering.DistributedLDAModel
-
Log likelihood of the observed tokens in the training set,
given the current parameter estimates:
log P(docs | topics, topic distributions for docs, Dirichlet hyperparameters)
Notes:
- This excludes the prior; for that, use logPrior
- trainOn(DStream<Vector>) - 类 中的方法org.apache.spark.mllib.clustering.StreamingKMeans
-
Update the clustering model by training on batches of data from a DStream.
- trainOn(JavaDStream<Vector>) - 类 中的方法org.apache.spark.mllib.clustering.StreamingKMeans
-
Java-friendly version of trainOn.
- trainOn(DStream<LabeledPoint>) - 类 中的方法org.apache.spark.mllib.regression.StreamingLinearAlgorithm
-
Update the model by training on batches of data from a DStream.
- trainOn(JavaDStream<LabeledPoint>) - 类 中的方法org.apache.spark.mllib.regression.StreamingLinearAlgorithm
-
Java-friendly version of trainOn.
- trainRatio() - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplit
-
- trainRatio() - 类 中的静态方法org.apache.spark.ml.tuning.TrainValidationSplitModel
-
- trainRegressor(RDD<LabeledPoint>, Map<Object, Object>, String, int, int) - 类 中的静态方法org.apache.spark.mllib.tree.DecisionTree
-
Method to train a decision tree model for regression.
- trainRegressor(JavaRDD<LabeledPoint>, Map<Integer, Integer>, String, int, int) - 类 中的静态方法org.apache.spark.mllib.tree.DecisionTree
-
Java-friendly API for org.apache.spark.mllib.tree.DecisionTree.trainRegressor
- trainRegressor(RDD<LabeledPoint>, Strategy, int, String, int) - 类 中的静态方法org.apache.spark.mllib.tree.RandomForest
-
Method to train a decision tree model for regression.
- trainRegressor(RDD<LabeledPoint>, Map<Object, Object>, int, String, String, int, int, int) - 类 中的静态方法org.apache.spark.mllib.tree.RandomForest
-
Method to train a decision tree model for regression.
- trainRegressor(JavaRDD<LabeledPoint>, Map<Integer, Integer>, int, String, String, int, int, int) - 类 中的静态方法org.apache.spark.mllib.tree.RandomForest
-
Java-friendly API for org.apache.spark.mllib.tree.RandomForest.trainRegressor
- TrainValidationSplit - org.apache.spark.ml.tuning中的类
-
Validation for hyper-parameter tuning.
- TrainValidationSplit(String) - 类 的构造器org.apache.spark.ml.tuning.TrainValidationSplit
-
- TrainValidationSplit() - 类 的构造器org.apache.spark.ml.tuning.TrainValidationSplit
-
- TrainValidationSplitModel - org.apache.spark.ml.tuning中的类
-
Model from train validation split.
- TrainValidationSplitModel.TrainValidationSplitModelWriter - org.apache.spark.ml.tuning中的类
-
Writer for TrainValidationSplitModel.
- transferred() - 类 中的方法org.apache.spark.storage.ReadableChannelFileRegion
-
- transferTo(WritableByteChannel, long) - 类 中的方法org.apache.spark.storage.ReadableChannelFileRegion
-
- transform(Function1<Try<T>, Try<S>>, ExecutionContext) - 类 中的方法org.apache.spark.ComplexFutureAction
-
- transform(Function1<Try<T>, Try<S>>, ExecutionContext) - 接口 中的方法org.apache.spark.FutureAction
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.classification.ClassificationModel
-
Transforms dataset by reading from featuresCol, and appending new columns as specified by
parameters:
- predicted labels as predictionCol of type Double
- raw predictions (confidences) as rawPredictionCol of type Vector.
- transform(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- transform(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- transform(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- transform(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- transform(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- transform(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- transform(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- transform(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- transform(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- transform(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- transform(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- transform(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- transform(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- transform(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- transform(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- transform(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- transform(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- transform(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- transform(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- transform(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- transform(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- transform(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- transform(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- transform(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.classification.OneVsRestModel
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
Transforms dataset by reading from featuresCol, and appending new columns as specified by
parameters:
- predicted labels as predictionCol of type Double
- raw predictions (confidences) as rawPredictionCol of type Vector
- probability of each class as probabilityCol of type Vector.
- transform(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- transform(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- transform(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- transform(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
- transform(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- transform(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- transform(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- transform(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.clustering.KMeansModel
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.clustering.LDAModel
-
Transforms the input dataset.
- transform(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- transform(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- transform(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- transform(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.Binarizer
-
- transform(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
-
- transform(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
-
- transform(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
-
- transform(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.Bucketizer
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.ColumnPruner
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.CountVectorizerModel
-
- transform(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.feature.DCT
-
- transform(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.feature.DCT
-
- transform(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.feature.DCT
-
- transform(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.feature.DCT
-
- transform(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.feature.ElementwiseProduct
-
- transform(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.feature.ElementwiseProduct
-
- transform(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.feature.ElementwiseProduct
-
- transform(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.feature.ElementwiseProduct
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.FeatureHasher
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.HashingTF
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.IDFModel
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.ImputerModel
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.IndexToString
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.Interaction
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.MaxAbsScalerModel
-
- transform(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSHModel
-
- transform(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSHModel
-
- transform(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSHModel
-
- transform(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSHModel
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.MinMaxScalerModel
-
- transform(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.feature.NGram
-
- transform(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.feature.NGram
-
- transform(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.feature.NGram
-
- transform(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.feature.NGram
-
- transform(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.feature.Normalizer
-
- transform(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.feature.Normalizer
-
- transform(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.feature.Normalizer
-
- transform(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.feature.Normalizer
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.OneHotEncoder
-
已过时。
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.OneHotEncoderModel
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.PCAModel
-
Transform a vector by computed Principal Components.
- transform(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.feature.PolynomialExpansion
-
- transform(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.feature.PolynomialExpansion
-
- transform(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.feature.PolynomialExpansion
-
- transform(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.feature.PolynomialExpansion
-
- transform(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.feature.RegexTokenizer
-
- transform(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.feature.RegexTokenizer
-
- transform(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.feature.RegexTokenizer
-
- transform(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.feature.RegexTokenizer
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.RFormulaModel
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.SQLTransformer
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.StandardScalerModel
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.StopWordsRemover
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.StringIndexerModel
-
- transform(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.feature.Tokenizer
-
- transform(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.feature.Tokenizer
-
- transform(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.feature.Tokenizer
-
- transform(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.feature.Tokenizer
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.VectorAssembler
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.VectorAttributeRewriter
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.VectorIndexerModel
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.VectorSizeHint
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.VectorSlicer
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.feature.Word2VecModel
-
Transform a sentence column to a vector column to represent the whole sentence.
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.fpm.FPGrowthModel
-
The transform method first generates the association rules according to the frequent itemsets.
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.PipelineModel
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.PredictionModel
-
Transforms dataset by reading from featuresCol, calling predict, and storing
the predictions as a new column predictionCol.
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.recommendation.ALSModel
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- transform(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- transform(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- transform(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- transform(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.regression.IsotonicRegressionModel
-
- transform(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- transform(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- transform(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- transform(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- transform(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- transform(Dataset<?>, ParamMap) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- transform(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- transform(Dataset<?>) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- transform(Dataset<?>, ParamPair<?>, ParamPair<?>...) - 类 中的方法org.apache.spark.ml.Transformer
-
Transforms the dataset with optional parameters
- transform(Dataset<?>, ParamPair<?>, Seq<ParamPair<?>>) - 类 中的方法org.apache.spark.ml.Transformer
-
Transforms the dataset with optional parameters
- transform(Dataset<?>, ParamMap) - 类 中的方法org.apache.spark.ml.Transformer
-
Transforms the dataset with provided parameter map as additional parameters.
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.Transformer
-
Transforms the input dataset.
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.tuning.CrossValidatorModel
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.tuning.TrainValidationSplitModel
-
- transform(Dataset<?>) - 类 中的方法org.apache.spark.ml.UnaryTransformer
-
- transform(Vector) - 类 中的方法org.apache.spark.mllib.feature.ChiSqSelectorModel
-
Applies transformation on a vector.
- transform(Vector) - 类 中的方法org.apache.spark.mllib.feature.ElementwiseProduct
-
Does the hadamard product transformation.
- transform(Iterable<?>) - 类 中的方法org.apache.spark.mllib.feature.HashingTF
-
Transforms the input document into a sparse term frequency vector.
- transform(Iterable<?>) - 类 中的方法org.apache.spark.mllib.feature.HashingTF
-
Transforms the input document into a sparse term frequency vector (Java version).
- transform(RDD<D>) - 类 中的方法org.apache.spark.mllib.feature.HashingTF
-
Transforms the input document to term frequency vectors.
- transform(JavaRDD<D>) - 类 中的方法org.apache.spark.mllib.feature.HashingTF
-
Transforms the input document to term frequency vectors (Java version).
- transform(RDD<Vector>) - 类 中的方法org.apache.spark.mllib.feature.IDFModel
-
Transforms term frequency (TF) vectors to TF-IDF vectors.
- transform(Vector) - 类 中的方法org.apache.spark.mllib.feature.IDFModel
-
Transforms a term frequency (TF) vector to a TF-IDF vector
- transform(JavaRDD<Vector>) - 类 中的方法org.apache.spark.mllib.feature.IDFModel
-
Transforms term frequency (TF) vectors to TF-IDF vectors (Java version).
- transform(Vector) - 类 中的方法org.apache.spark.mllib.feature.Normalizer
-
Applies unit length normalization on a vector.
- transform(Vector) - 类 中的方法org.apache.spark.mllib.feature.PCAModel
-
Transform a vector by computed Principal Components.
- transform(Vector) - 类 中的方法org.apache.spark.mllib.feature.StandardScalerModel
-
Applies standardization transformation on a vector.
- transform(Vector) - 接口 中的方法org.apache.spark.mllib.feature.VectorTransformer
-
Applies transformation on a vector.
- transform(RDD<Vector>) - 接口 中的方法org.apache.spark.mllib.feature.VectorTransformer
-
Applies transformation on an RDD[Vector].
- transform(JavaRDD<Vector>) - 接口 中的方法org.apache.spark.mllib.feature.VectorTransformer
-
Applies transformation on a JavaRDD[Vector].
- transform(String) - 类 中的方法org.apache.spark.mllib.feature.Word2VecModel
-
Transforms a word to its vector representation
- transform(Function1<Try<T>, Try<S>>, ExecutionContext) - 类 中的方法org.apache.spark.SimpleFutureAction
-
- transform(Function1<Dataset<T>, Dataset<U>>) - 类 中的方法org.apache.spark.sql.Dataset
-
Concise syntax for chaining custom transformations.
- transform(PartialFunction<BaseType, BaseType>) - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- transform(PartialFunction<BaseType, BaseType>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- transform(PartialFunction<BaseType, BaseType>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- transform(PartialFunction<BaseType, BaseType>) - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- transform(Function<R, JavaRDD<U>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- transform(Function2<R, Time, JavaRDD<U>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- transform(Function<R, JavaRDD<U>>) - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return a new DStream in which each RDD is generated by applying a function
on each RDD of 'this' DStream.
- transform(Function2<R, Time, JavaRDD<U>>) - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return a new DStream in which each RDD is generated by applying a function
on each RDD of 'this' DStream.
- transform(Function<R, JavaRDD<U>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- transform(Function2<R, Time, JavaRDD<U>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- transform(Function<R, JavaRDD<U>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- transform(Function2<R, Time, JavaRDD<U>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- transform(Function<R, JavaRDD<U>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- transform(Function2<R, Time, JavaRDD<U>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- transform(Function<R, JavaRDD<U>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- transform(Function2<R, Time, JavaRDD<U>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- transform(Function<R, JavaRDD<U>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- transform(Function2<R, Time, JavaRDD<U>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- transform(List<JavaDStream<?>>, Function2<List<JavaRDD<?>>, Time, JavaRDD<T>>) - 类 中的方法org.apache.spark.streaming.api.java.JavaStreamingContext
-
Create a new DStream in which each RDD is generated by applying a function on RDDs of
the DStreams.
- transform(Function1<RDD<T>, RDD<U>>, ClassTag<U>) - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Return a new DStream in which each RDD is generated by applying a function
on each RDD of 'this' DStream.
- transform(Function2<RDD<T>, Time, RDD<U>>, ClassTag<U>) - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Return a new DStream in which each RDD is generated by applying a function
on each RDD of 'this' DStream.
- transform(Seq<DStream<?>>, Function2<Seq<RDD<?>>, Time, RDD<T>>, ClassTag<T>) - 类 中的方法org.apache.spark.streaming.StreamingContext
-
Create a new DStream in which each RDD is generated by applying a function on RDDs of
the DStreams.
- transformAllExpressions(PartialFunction<Expression, Expression>) - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- transformAllExpressions(PartialFunction<Expression, Expression>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- transformAllExpressions(PartialFunction<Expression, Expression>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- transformAllExpressions(PartialFunction<Expression, Expression>) - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- transformDown(PartialFunction<BaseType, BaseType>) - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- transformDown(PartialFunction<BaseType, BaseType>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- transformDown(PartialFunction<BaseType, BaseType>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- transformDown(PartialFunction<BaseType, BaseType>) - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- Transformer - org.apache.spark.ml中的类
-
:: DeveloperApi ::
Abstract class for transformers that transform one dataset into another.
- Transformer() - 类 的构造器org.apache.spark.ml.Transformer
-
- transformExpressions(PartialFunction<Expression, Expression>) - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- transformExpressions(PartialFunction<Expression, Expression>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- transformExpressions(PartialFunction<Expression, Expression>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- transformExpressions(PartialFunction<Expression, Expression>) - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- transformExpressionsDown(PartialFunction<Expression, Expression>) - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- transformExpressionsDown(PartialFunction<Expression, Expression>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- transformExpressionsDown(PartialFunction<Expression, Expression>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- transformExpressionsDown(PartialFunction<Expression, Expression>) - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- transformExpressionsUp(PartialFunction<Expression, Expression>) - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- transformExpressionsUp(PartialFunction<Expression, Expression>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- transformExpressionsUp(PartialFunction<Expression, Expression>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- transformExpressionsUp(PartialFunction<Expression, Expression>) - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- transformOutputColumnSchema(StructField, String, boolean, boolean) - 类 中的静态方法org.apache.spark.ml.feature.OneHotEncoderCommon
-
Prepares the StructField with proper metadata for OneHotEncoder's output column.
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassificationModel
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.classification.DecisionTreeClassifier
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassificationModel
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.classification.GBTClassifier
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVC
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.classification.LinearSVCModel
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegression
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.classification.LogisticRegressionModel
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassificationModel
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.classification.MultilayerPerceptronClassifier
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayes
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.classification.NaiveBayesModel
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.classification.OneVsRest
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.classification.OneVsRestModel
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.classification.ProbabilisticClassificationModel
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.classification.RandomForestClassifier
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.clustering.BisectingKMeans
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.clustering.BisectingKMeansModel
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.clustering.DistributedLDAModel
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.clustering.GaussianMixture
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.clustering.GaussianMixtureModel
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.clustering.KMeans
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.clustering.KMeansModel
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.clustering.LDA
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.clustering.LDAModel
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.clustering.LocalLDAModel
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.Binarizer
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.BucketedRandomProjectionLSH
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.feature.BucketedRandomProjectionLSHModel
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.Bucketizer
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.ChiSqSelector
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.ChiSqSelectorModel
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.ColumnPruner
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.CountVectorizer
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.CountVectorizerModel
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.feature.DCT
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.feature.ElementwiseProduct
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.FeatureHasher
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.HashingTF
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.IDF
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.IDFModel
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.Imputer
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.ImputerModel
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.IndexToString
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.Interaction
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.MaxAbsScaler
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.MaxAbsScalerModel
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.MinHashLSH
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.feature.MinHashLSHModel
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.MinMaxScaler
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.MinMaxScalerModel
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.feature.NGram
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.feature.Normalizer
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.OneHotEncoder
-
已过时。
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.OneHotEncoderEstimator
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.OneHotEncoderModel
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.PCA
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.PCAModel
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.feature.PolynomialExpansion
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.QuantileDiscretizer
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.feature.RegexTokenizer
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.RFormula
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.RFormulaModel
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.SQLTransformer
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.StandardScaler
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.StandardScalerModel
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.StopWordsRemover
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.StringIndexer
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.StringIndexerModel
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.feature.Tokenizer
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.VectorAssembler
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.VectorAttributeRewriter
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.VectorIndexer
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.VectorIndexerModel
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.VectorSizeHint
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.VectorSlicer
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.Word2Vec
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.feature.Word2VecModel
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.fpm.FPGrowth
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.fpm.FPGrowthModel
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.Pipeline
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.PipelineModel
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.PipelineStage
-
:: DeveloperApi ::
Check transform validity and derive the output schema from the input schema.
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.PredictionModel
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.Predictor
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.recommendation.ALS
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.recommendation.ALSModel
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.regression.AFTSurvivalRegression
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.regression.AFTSurvivalRegressionModel
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressionModel
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.regression.DecisionTreeRegressor
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressionModel
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.regression.GBTRegressor
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegression
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.regression.GeneralizedLinearRegressionModel
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.regression.IsotonicRegression
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.regression.IsotonicRegressionModel
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegression
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.regression.LinearRegressionModel
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- transformSchema(StructType) - 类 中的静态方法org.apache.spark.ml.regression.RandomForestRegressor
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.tuning.CrossValidator
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.tuning.CrossValidatorModel
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.tuning.TrainValidationSplit
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.tuning.TrainValidationSplitModel
-
- transformSchema(StructType) - 类 中的方法org.apache.spark.ml.UnaryTransformer
-
- transformToPair(Function<R, JavaPairRDD<K2, V2>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- transformToPair(Function2<R, Time, JavaPairRDD<K2, V2>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- transformToPair(Function<R, JavaPairRDD<K2, V2>>) - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return a new DStream in which each RDD is generated by applying a function
on each RDD of 'this' DStream.
- transformToPair(Function2<R, Time, JavaPairRDD<K2, V2>>) - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return a new DStream in which each RDD is generated by applying a function
on each RDD of 'this' DStream.
- transformToPair(Function<R, JavaPairRDD<K2, V2>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- transformToPair(Function2<R, Time, JavaPairRDD<K2, V2>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- transformToPair(Function<R, JavaPairRDD<K2, V2>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- transformToPair(Function2<R, Time, JavaPairRDD<K2, V2>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- transformToPair(Function<R, JavaPairRDD<K2, V2>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- transformToPair(Function2<R, Time, JavaPairRDD<K2, V2>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- transformToPair(Function<R, JavaPairRDD<K2, V2>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- transformToPair(Function2<R, Time, JavaPairRDD<K2, V2>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- transformToPair(Function<R, JavaPairRDD<K2, V2>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- transformToPair(Function2<R, Time, JavaPairRDD<K2, V2>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- transformToPair(List<JavaDStream<?>>, Function2<List<JavaRDD<?>>, Time, JavaPairRDD<K, V>>) - 类 中的方法org.apache.spark.streaming.api.java.JavaStreamingContext
-
Create a new DStream in which each RDD is generated by applying a function on RDDs of
the DStreams.
- transformUp(PartialFunction<BaseType, BaseType>) - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- transformUp(PartialFunction<BaseType, BaseType>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- transformUp(PartialFunction<BaseType, BaseType>) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- transformUp(PartialFunction<BaseType, BaseType>) - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- transformWith(Function1<Try<T>, Future<S>>, ExecutionContext) - 类 中的方法org.apache.spark.ComplexFutureAction
-
- transformWith(Function1<Try<T>, Future<S>>, ExecutionContext) - 接口 中的方法org.apache.spark.FutureAction
-
- transformWith(Function1<Try<T>, Future<S>>, ExecutionContext) - 类 中的方法org.apache.spark.SimpleFutureAction
-
- transformWith(JavaDStream<U>, Function3<R, JavaRDD<U>, Time, JavaRDD<W>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- transformWith(JavaPairDStream<K2, V2>, Function3<R, JavaPairRDD<K2, V2>, Time, JavaRDD<W>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- transformWith(JavaDStream<U>, Function3<R, JavaRDD<U>, Time, JavaRDD<W>>) - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return a new DStream in which each RDD is generated by applying a function
on each RDD of 'this' DStream and 'other' DStream.
- transformWith(JavaPairDStream<K2, V2>, Function3<R, JavaPairRDD<K2, V2>, Time, JavaRDD<W>>) - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return a new DStream in which each RDD is generated by applying a function
on each RDD of 'this' DStream and 'other' DStream.
- transformWith(JavaDStream<U>, Function3<R, JavaRDD<U>, Time, JavaRDD<W>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- transformWith(JavaPairDStream<K2, V2>, Function3<R, JavaPairRDD<K2, V2>, Time, JavaRDD<W>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- transformWith(JavaDStream<U>, Function3<R, JavaRDD<U>, Time, JavaRDD<W>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- transformWith(JavaPairDStream<K2, V2>, Function3<R, JavaPairRDD<K2, V2>, Time, JavaRDD<W>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- transformWith(JavaDStream<U>, Function3<R, JavaRDD<U>, Time, JavaRDD<W>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- transformWith(JavaPairDStream<K2, V2>, Function3<R, JavaPairRDD<K2, V2>, Time, JavaRDD<W>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- transformWith(JavaDStream<U>, Function3<R, JavaRDD<U>, Time, JavaRDD<W>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- transformWith(JavaPairDStream<K2, V2>, Function3<R, JavaPairRDD<K2, V2>, Time, JavaRDD<W>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- transformWith(JavaDStream<U>, Function3<R, JavaRDD<U>, Time, JavaRDD<W>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- transformWith(JavaPairDStream<K2, V2>, Function3<R, JavaPairRDD<K2, V2>, Time, JavaRDD<W>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- transformWith(DStream<U>, Function2<RDD<T>, RDD<U>, RDD<V>>, ClassTag<U>, ClassTag<V>) - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Return a new DStream in which each RDD is generated by applying a function
on each RDD of 'this' DStream and 'other' DStream.
- transformWith(DStream<U>, Function3<RDD<T>, RDD<U>, Time, RDD<V>>, ClassTag<U>, ClassTag<V>) - 类 中的方法org.apache.spark.streaming.dstream.DStream
-
Return a new DStream in which each RDD is generated by applying a function
on each RDD of 'this' DStream and 'other' DStream.
- transformWithToPair(JavaDStream<U>, Function3<R, JavaRDD<U>, Time, JavaPairRDD<K2, V2>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- transformWithToPair(JavaPairDStream<K2, V2>, Function3<R, JavaPairRDD<K2, V2>, Time, JavaPairRDD<K3, V3>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaDStream
-
- transformWithToPair(JavaDStream<U>, Function3<R, JavaRDD<U>, Time, JavaPairRDD<K2, V2>>) - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return a new DStream in which each RDD is generated by applying a function
on each RDD of 'this' DStream and 'other' DStream.
- transformWithToPair(JavaPairDStream<K2, V2>, Function3<R, JavaPairRDD<K2, V2>, Time, JavaPairRDD<K3, V3>>) - 接口 中的方法org.apache.spark.streaming.api.java.JavaDStreamLike
-
Return a new DStream in which each RDD is generated by applying a function
on each RDD of 'this' DStream and 'other' DStream.
- transformWithToPair(JavaDStream<U>, Function3<R, JavaRDD<U>, Time, JavaPairRDD<K2, V2>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- transformWithToPair(JavaPairDStream<K2, V2>, Function3<R, JavaPairRDD<K2, V2>, Time, JavaPairRDD<K3, V3>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaInputDStream
-
- transformWithToPair(JavaDStream<U>, Function3<R, JavaRDD<U>, Time, JavaPairRDD<K2, V2>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- transformWithToPair(JavaPairDStream<K2, V2>, Function3<R, JavaPairRDD<K2, V2>, Time, JavaPairRDD<K3, V3>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairDStream
-
- transformWithToPair(JavaDStream<U>, Function3<R, JavaRDD<U>, Time, JavaPairRDD<K2, V2>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- transformWithToPair(JavaPairDStream<K2, V2>, Function3<R, JavaPairRDD<K2, V2>, Time, JavaPairRDD<K3, V3>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairInputDStream
-
- transformWithToPair(JavaDStream<U>, Function3<R, JavaRDD<U>, Time, JavaPairRDD<K2, V2>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- transformWithToPair(JavaPairDStream<K2, V2>, Function3<R, JavaPairRDD<K2, V2>, Time, JavaPairRDD<K3, V3>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaPairReceiverInputDStream
-
- transformWithToPair(JavaDStream<U>, Function3<R, JavaRDD<U>, Time, JavaPairRDD<K2, V2>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- transformWithToPair(JavaPairDStream<K2, V2>, Function3<R, JavaPairRDD<K2, V2>, Time, JavaPairRDD<K3, V3>>) - 类 中的静态方法org.apache.spark.streaming.api.java.JavaReceiverInputDStream
-
- translate(Column, String, String) - 类 中的静态方法org.apache.spark.sql.functions
-
Translate any character in the src by a character in replaceString.
- transpose() - 类 中的方法org.apache.spark.ml.linalg.DenseMatrix
-
- transpose() - 接口 中的方法org.apache.spark.ml.linalg.Matrix
-
Transpose the Matrix.
- transpose() - 类 中的方法org.apache.spark.ml.linalg.SparseMatrix
-
- transpose() - 类 中的方法org.apache.spark.mllib.linalg.DenseMatrix
-
- transpose() - 类 中的方法org.apache.spark.mllib.linalg.distributed.BlockMatrix
-
Transpose this BlockMatrix.
- transpose() - 类 中的方法org.apache.spark.mllib.linalg.distributed.CoordinateMatrix
-
Transposes this CoordinateMatrix.
- transpose() - 接口 中的方法org.apache.spark.mllib.linalg.Matrix
-
Transpose the Matrix.
- transpose() - 类 中的方法org.apache.spark.mllib.linalg.SparseMatrix
-
- transpose(Function1<A, GenTraversableOnce<B>>) - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- treeAggregate(U, Function2<U, T, U>, Function2<U, U, U>, int) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- treeAggregate(U, Function2<U, T, U>, Function2<U, U, U>) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- treeAggregate(U, Function2<U, T, U>, Function2<U, U, U>, int) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- treeAggregate(U, Function2<U, T, U>, Function2<U, U, U>) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- treeAggregate(U, Function2<U, T, U>, Function2<U, U, U>, int) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- treeAggregate(U, Function2<U, T, U>, Function2<U, U, U>) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- treeAggregate(U, Function2<U, T, U>, Function2<U, U, U>, int) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Aggregates the elements of this RDD in a multi-level tree pattern.
- treeAggregate(U, Function2<U, T, U>, Function2<U, U, U>) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
org.apache.spark.api.java.JavaRDDLike.treeAggregate with suggested depth 2.
- treeAggregate(U, Function2<U, T, U>, Function2<U, U, U>, int, ClassTag<U>) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- treeAggregate(U, Function2<U, T, U>, Function2<U, U, U>, int, ClassTag<U>) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- treeAggregate(U, Function2<U, T, U>, Function2<U, U, U>, int, ClassTag<U>) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- treeAggregate(U, Function2<U, T, U>, Function2<U, U, U>, int, ClassTag<U>) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- treeAggregate(U, Function2<U, T, U>, Function2<U, U, U>, int, ClassTag<U>) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- treeAggregate(U, Function2<U, T, U>, Function2<U, U, U>, int, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- treeAggregate(U, Function2<U, T, U>, Function2<U, U, U>, int, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- treeAggregate(U, Function2<U, T, U>, Function2<U, U, U>, int, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- treeAggregate(U, Function2<U, T, U>, Function2<U, U, U>, int, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- treeAggregate(U, Function2<U, T, U>, Function2<U, U, U>, int, ClassTag<U>) - 类 中的方法org.apache.spark.rdd.RDD
-
Aggregates the elements of this RDD in a multi-level tree pattern.
- treeAggregate(U, Function2<U, T, U>, Function2<U, U, U>, int, ClassTag<U>) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- treeAggregate$default$4(U) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- treeAggregate$default$4(U) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- treeAggregate$default$4(U) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- treeAggregate$default$4(U) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- treeAggregate$default$4(U) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- treeAggregate$default$4(U) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- treeAggregate$default$4(U) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- treeAggregate$default$4(U) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- treeAggregate$default$4(U) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- treeAggregate$default$4(U) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- treeID() - 类 中的方法org.apache.spark.ml.tree.EnsembleModelReadWrite.EnsembleNodeData
-
- treeId() - 类 中的方法org.apache.spark.mllib.tree.model.DecisionTreeModel.SaveLoadV1_0$.NodeData
-
- treeReduce(Function2<T, T, T>, int) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- treeReduce(Function2<T, T, T>) - 类 中的静态方法org.apache.spark.api.java.JavaDoubleRDD
-
- treeReduce(Function2<T, T, T>, int) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- treeReduce(Function2<T, T, T>) - 类 中的静态方法org.apache.spark.api.java.JavaPairRDD
-
- treeReduce(Function2<T, T, T>, int) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- treeReduce(Function2<T, T, T>) - 类 中的静态方法org.apache.spark.api.java.JavaRDD
-
- treeReduce(Function2<T, T, T>, int) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
Reduces the elements of this RDD in a multi-level tree pattern.
- treeReduce(Function2<T, T, T>) - 接口 中的方法org.apache.spark.api.java.JavaRDDLike
-
org.apache.spark.api.java.JavaRDDLike.treeReduce with suggested depth 2.
- treeReduce(Function2<T, T, T>, int) - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- treeReduce(Function2<T, T, T>, int) - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- treeReduce(Function2<T, T, T>, int) - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- treeReduce(Function2<T, T, T>, int) - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- treeReduce(Function2<T, T, T>, int) - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- treeReduce(Function2<T, T, T>, int) - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- treeReduce(Function2<T, T, T>, int) - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- treeReduce(Function2<T, T, T>, int) - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- treeReduce(Function2<T, T, T>, int) - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- treeReduce(Function2<T, T, T>, int) - 类 中的方法org.apache.spark.rdd.RDD
-
Reduces the elements of this RDD in a multi-level tree pattern.
- treeReduce(Function2<T, T, T>, int) - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- treeReduce$default$2() - 类 中的静态方法org.apache.spark.api.r.RRDD
-
- treeReduce$default$2() - 类 中的静态方法org.apache.spark.graphx.EdgeRDD
-
- treeReduce$default$2() - 类 中的静态方法org.apache.spark.graphx.impl.EdgeRDDImpl
-
- treeReduce$default$2() - 类 中的静态方法org.apache.spark.graphx.impl.VertexRDDImpl
-
- treeReduce$default$2() - 类 中的静态方法org.apache.spark.graphx.VertexRDD
-
- treeReduce$default$2() - 类 中的静态方法org.apache.spark.rdd.HadoopRDD
-
- treeReduce$default$2() - 类 中的静态方法org.apache.spark.rdd.JdbcRDD
-
- treeReduce$default$2() - 类 中的静态方法org.apache.spark.rdd.NewHadoopRDD
-
- treeReduce$default$2() - 类 中的静态方法org.apache.spark.rdd.PartitionPruningRDD
-
- treeReduce$default$2() - 类 中的静态方法org.apache.spark.rdd.UnionRDD
-
- trees() - 类 中的方法org.apache.spark.ml.classification.GBTClassificationModel
-
- trees() - 类 中的方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- trees() - 类 中的方法org.apache.spark.ml.regression.GBTRegressionModel
-
- trees() - 类 中的方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- trees() - 类 中的方法org.apache.spark.mllib.tree.model.GradientBoostedTreesModel
-
- trees() - 类 中的方法org.apache.spark.mllib.tree.model.RandomForestModel
-
- treeStrategy() - 类 中的方法org.apache.spark.mllib.tree.configuration.BoostingStrategy
-
- treeString() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- treeString(boolean, boolean) - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- treeString() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- treeString(boolean, boolean) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- treeString() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- treeString(boolean, boolean) - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- treeString() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- treeString(boolean, boolean) - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- treeString() - 类 中的方法org.apache.spark.sql.types.StructType
-
- treeString$default$2() - 类 中的静态方法org.apache.spark.sql.hive.execution.CreateHiveTableAsSelectCommand
-
- treeString$default$2() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveDirCommand
-
- treeString$default$2() - 类 中的静态方法org.apache.spark.sql.hive.execution.InsertIntoHiveTable
-
- treeString$default$2() - 类 中的静态方法org.apache.spark.sql.hive.execution.ScriptTransformationExec
-
- treeWeights() - 类 中的方法org.apache.spark.ml.classification.GBTClassificationModel
-
- treeWeights() - 类 中的方法org.apache.spark.ml.classification.RandomForestClassificationModel
-
- treeWeights() - 类 中的方法org.apache.spark.ml.regression.GBTRegressionModel
-
- treeWeights() - 类 中的方法org.apache.spark.ml.regression.RandomForestRegressionModel
-
- treeWeights() - 类 中的方法org.apache.spark.mllib.tree.model.GradientBoostedTreesModel
-
- triangleCount() - 类 中的方法org.apache.spark.graphx.GraphOps
-
Compute the number of triangles passing through each vertex.
- TriangleCount - org.apache.spark.graphx.lib中的类
-
Compute the number of triangles passing through each vertex.
- TriangleCount() - 类 的构造器org.apache.spark.graphx.lib.TriangleCount
-
- trigger(Trigger) - 类 中的方法org.apache.spark.sql.streaming.DataStreamWriter
-
Set the trigger for the stream query.
- Trigger - org.apache.spark.sql.streaming中的类
-
Policy used to indicate how often results should be produced by a [[StreamingQuery]].
- Trigger() - 类 的构造器org.apache.spark.sql.streaming.Trigger
-
- TriggerThreadDump$() - 类 的构造器org.apache.spark.storage.BlockManagerMessages.TriggerThreadDump$
-
- trim(Column) - 类 中的静态方法org.apache.spark.sql.functions
-
Trim the spaces from both ends for the specified string column.
- trim(Column, String) - 类 中的静态方法org.apache.spark.sql.functions
-
Trim the specified character from both ends for the specified string column.
- TrimHorizon() - 类 的构造器org.apache.spark.streaming.kinesis.KinesisInitialPositions.TrimHorizon
-
- TripletFields - org.apache.spark.graphx中的类
-
Represents a subset of the fields of an [[EdgeTriplet]] or [[EdgeContext]].
- TripletFields() - 类 的构造器org.apache.spark.graphx.TripletFields
-
Constructs a default TripletFields in which all fields are included.
- TripletFields(boolean, boolean, boolean) - 类 的构造器org.apache.spark.graphx.TripletFields
-
- triplets() - 类 中的方法org.apache.spark.graphx.Graph
-
An RDD containing the edge triplets, which are edges along with the vertex data associated with
the adjacent vertices.
- triplets() - 类 中的方法org.apache.spark.graphx.impl.GraphImpl
-
Return an RDD that brings edges together with their source and destination vertices.
- truePositiveRate(double) - 类 中的方法org.apache.spark.mllib.evaluation.MulticlassMetrics
-
Returns true positive rate for a given label (category)
- truePositiveRateByLabel() - 接口 中的方法org.apache.spark.ml.classification.LogisticRegressionSummary
-
Returns true positive rate for each label (category).
- trunc(Column, String) - 类 中的静态方法org.apache.spark.sql.functions
-
Returns date truncated to the unit specified by the format.
- truncatedString(Seq<T>, String, String, String, int) - 类 中的静态方法org.apache.spark.util.Utils
-
Format a sequence with semantics similar to calling .mkString().
- truncatedString(Seq<T>, String) - 类 中的静态方法org.apache.spark.util.Utils
-
Shorthand for calling truncatedString() without start or end strings.
- tryLog(Function0<T>) - 类 中的静态方法org.apache.spark.util.Utils
-
Executes the given block in a Try, logging any uncaught exceptions.
- tryLogNonFatalError(Function0<BoxedUnit>) - 类 中的静态方法org.apache.spark.util.Utils
-
Executes the given block.
- tryOrExit(Function0<BoxedUnit>) - 类 中的静态方法org.apache.spark.util.Utils
-
Execute a block of code that evaluates to Unit, forwarding any uncaught exceptions to the
default UncaughtExceptionHandler
NOTE: This method is to be called by the spark-started JVM process.
- tryOrIOException(Function0<T>) - 类 中的静态方法org.apache.spark.util.Utils
-
Execute a block of code that returns a value, re-throwing any non-fatal uncaught
exceptions as IOException.
- tryOrStopSparkContext(SparkContext, Function0<BoxedUnit>) - 类 中的静态方法org.apache.spark.util.Utils
-
Execute a block of code that evaluates to Unit, stop SparkContext if there is any uncaught
exception
NOTE: This method is to be called by the driver-side components to avoid stopping the
user-started JVM process completely; in contrast, tryOrExit is to be called in the
spark-started JVM process .
- tryRecoverFromCheckpoint(String) - 类 中的方法org.apache.spark.streaming.StreamingContextPythonHelper
-
This is a private method only for Python to implement getOrCreate.
- tryWithResource(Function0<R>, Function1<R, T>) - 类 中的静态方法org.apache.spark.util.Utils
-
- tryWithSafeFinally(Function0<T>, Function0<BoxedUnit>) - 类 中的静态方法org.apache.spark.util.Utils
-
Execute a block of code, then a finally block, but if exceptions happen in
the finally block, do not suppress the original exception.
- tryWithSafeFinallyAndFailureCallbacks(Function0<T>, Function0<BoxedUnit>, Function0<BoxedUnit>) - 类 中的静态方法org.apache.spark.util.Utils
-
Execute a block of code and call the failure callbacks in the catch block.
- tuple(Encoder<T1>, Encoder<T2>) - 类 中的静态方法org.apache.spark.sql.Encoders
-
An encoder for 2-ary tuples.
- tuple(Encoder<T1>, Encoder<T2>, Encoder<T3>) - 类 中的静态方法org.apache.spark.sql.Encoders
-
An encoder for 3-ary tuples.
- tuple(Encoder<T1>, Encoder<T2>, Encoder<T3>, Encoder<T4>) - 类 中的静态方法org.apache.spark.sql.Encoders
-
An encoder for 4-ary tuples.
- tuple(Encoder<T1>, Encoder<T2>, Encoder<T3>, Encoder<T4>, Encoder<T5>) - 类 中的静态方法org.apache.spark.sql.Encoders
-
An encoder for 5-ary tuples.
- tValues() - 类 中的方法org.apache.spark.ml.regression.GeneralizedLinearRegressionTrainingSummary
-
T-statistic of estimated coefficients and intercept.
- tValues() - 类 中的方法org.apache.spark.ml.regression.LinearRegressionSummary
-
T-statistic of estimated coefficients and intercept.
- Tweedie$() - 类 的构造器org.apache.spark.ml.regression.GeneralizedLinearRegression.Tweedie$
-
- TYPE() - 类 中的静态方法org.apache.spark.ml.attribute.AttributeKeys
-
- type() - 类 中的方法org.apache.spark.util.kvstore.KVTypeInfo
-
- typed - org.apache.spark.sql.expressions.javalang中的类
-
:: Experimental ::
Type-safe functions available for
Dataset operations in Java.
- typed() - 类 的构造器org.apache.spark.sql.expressions.javalang.typed
-
- typed - org.apache.spark.sql.expressions.scalalang中的类
-
:: Experimental ::
Type-safe functions available for Dataset operations in Scala.
- typed() - 类 的构造器org.apache.spark.sql.expressions.scalalang.typed
-
- TypedColumn<T,U> - org.apache.spark.sql中的类
-
A
Column where an
Encoder has been given for the expected input and return type.
- TypedColumn(Expression, ExpressionEncoder<U>) - 类 的构造器org.apache.spark.sql.TypedColumn
-
- typedLit(T, TypeTags.TypeTag<T>) - 类 中的静态方法org.apache.spark.sql.functions
-
Creates a
Column of literal value.
- typeInfoConversions(DataType) - 类 中的静态方法org.apache.spark.sql.hive.execution.HiveScriptIOSchema
-
- typeName() - 类 中的方法org.apache.spark.mllib.linalg.VectorUDT
-
- typeName() - 类 中的静态方法org.apache.spark.sql.types.ArrayType
-
- typeName() - 类 中的静态方法org.apache.spark.sql.types.BinaryType
-
- typeName() - 类 中的静态方法org.apache.spark.sql.types.BooleanType
-
- typeName() - 类 中的静态方法org.apache.spark.sql.types.ByteType
-
- typeName() - 类 中的静态方法org.apache.spark.sql.types.CalendarIntervalType
-
- typeName() - 类 中的静态方法org.apache.spark.sql.types.CharType
-
- typeName() - 类 中的方法org.apache.spark.sql.types.DataType
-
Name of the type used in JSON serialization.
- typeName() - 类 中的静态方法org.apache.spark.sql.types.DateType
-
- typeName() - 类 中的方法org.apache.spark.sql.types.DecimalType
-
- typeName() - 类 中的静态方法org.apache.spark.sql.types.DoubleType
-
- typeName() - 类 中的静态方法org.apache.spark.sql.types.FloatType
-
- typeName() - 类 中的静态方法org.apache.spark.sql.types.HiveStringType
-
- typeName() - 类 中的静态方法org.apache.spark.sql.types.IntegerType
-
- typeName() - 类 中的静态方法org.apache.spark.sql.types.LongType
-
- typeName() - 类 中的静态方法org.apache.spark.sql.types.MapType
-
- typeName() - 类 中的静态方法org.apache.spark.sql.types.NullType
-
- typeName() - 类 中的静态方法org.apache.spark.sql.types.NumericType
-
- typeName() - 类 中的静态方法org.apache.spark.sql.types.ObjectType
-
- typeName() - 类 中的静态方法org.apache.spark.sql.types.ShortType
-
- typeName() - 类 中的静态方法org.apache.spark.sql.types.StringType
-
- typeName() - 类 中的静态方法org.apache.spark.sql.types.StructType
-
- typeName() - 类 中的静态方法org.apache.spark.sql.types.TimestampType
-
- typeName() - 类 中的静态方法org.apache.spark.sql.types.VarcharType
-